You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2017/08/02 16:54:27 UTC

[01/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Repository: hbase
Updated Branches:
  refs/heads/branch-2 f92334299 -> 6440509d7


http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/BackupProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/BackupProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/BackupProtos.java
deleted file mode 100644
index 0271777..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/BackupProtos.java
+++ /dev/null
@@ -1,7013 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: Backup.proto
-
-package org.apache.hadoop.hbase.shaded.protobuf.generated;
-
-public final class BackupProtos {
-  private BackupProtos() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  /**
-   * Protobuf enum {@code hbase.pb.BackupType}
-   */
-  public enum BackupType
-      implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
-    /**
-     * <code>FULL = 0;</code>
-     */
-    FULL(0),
-    /**
-     * <code>INCREMENTAL = 1;</code>
-     */
-    INCREMENTAL(1),
-    ;
-
-    /**
-     * <code>FULL = 0;</code>
-     */
-    public static final int FULL_VALUE = 0;
-    /**
-     * <code>INCREMENTAL = 1;</code>
-     */
-    public static final int INCREMENTAL_VALUE = 1;
-
-
-    public final int getNumber() {
-      return value;
-    }
-
-    /**
-     * @deprecated Use {@link #forNumber(int)} instead.
-     */
-    @java.lang.Deprecated
-    public static BackupType valueOf(int value) {
-      return forNumber(value);
-    }
-
-    public static BackupType forNumber(int value) {
-      switch (value) {
-        case 0: return FULL;
-        case 1: return INCREMENTAL;
-        default: return null;
-      }
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<BackupType>
-        internalGetValueMap() {
-      return internalValueMap;
-    }
-    private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
-        BackupType> internalValueMap =
-          new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<BackupType>() {
-            public BackupType findValueByNumber(int number) {
-              return BackupType.forNumber(number);
-            }
-          };
-
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
-        getValueDescriptor() {
-      return getDescriptor().getValues().get(ordinal());
-    }
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.getDescriptor().getEnumTypes().get(0);
-    }
-
-    private static final BackupType[] VALUES = values();
-
-    public static BackupType valueOf(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-      if (desc.getType() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "EnumValueDescriptor is not for this type.");
-      }
-      return VALUES[desc.getIndex()];
-    }
-
-    private final int value;
-
-    private BackupType(int value) {
-      this.value = value;
-    }
-
-    // @@protoc_insertion_point(enum_scope:hbase.pb.BackupType)
-  }
-
-  public interface ServerTimestampOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:hbase.pb.ServerTimestamp)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-     */
-    boolean hasServerName();
-    /**
-     * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName();
-    /**
-     * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder();
-
-    /**
-     * <code>optional uint64 timestamp = 2;</code>
-     */
-    boolean hasTimestamp();
-    /**
-     * <code>optional uint64 timestamp = 2;</code>
-     */
-    long getTimestamp();
-  }
-  /**
-   * <pre>
-   **
-   * ServerTimestamp keeps last WAL roll time per Region Server
-   * </pre>
-   *
-   * Protobuf type {@code hbase.pb.ServerTimestamp}
-   */
-  public  static final class ServerTimestamp extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:hbase.pb.ServerTimestamp)
-      ServerTimestampOrBuilder {
-    // Use ServerTimestamp.newBuilder() to construct.
-    private ServerTimestamp(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private ServerTimestamp() {
-      timestamp_ = 0L;
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private ServerTimestamp(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = serverName_.toBuilder();
-              }
-              serverName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(serverName_);
-                serverName_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000001;
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              timestamp_ = input.readUInt64();
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_ServerTimestamp_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_ServerTimestamp_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.class, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int SERVER_NAME_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_;
-    /**
-     * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-     */
-    public boolean hasServerName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName() {
-      return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_;
-    }
-    /**
-     * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() {
-      return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_;
-    }
-
-    public static final int TIMESTAMP_FIELD_NUMBER = 2;
-    private long timestamp_;
-    /**
-     * <code>optional uint64 timestamp = 2;</code>
-     */
-    public boolean hasTimestamp() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional uint64 timestamp = 2;</code>
-     */
-    public long getTimestamp() {
-      return timestamp_;
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (hasServerName()) {
-        if (!getServerName().isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, getServerName());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeUInt64(2, timestamp_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, getServerName());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeUInt64Size(2, timestamp_);
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp other = (org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp) obj;
-
-      boolean result = true;
-      result = result && (hasServerName() == other.hasServerName());
-      if (hasServerName()) {
-        result = result && getServerName()
-            .equals(other.getServerName());
-      }
-      result = result && (hasTimestamp() == other.hasTimestamp());
-      if (hasTimestamp()) {
-        result = result && (getTimestamp()
-            == other.getTimestamp());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasServerName()) {
-        hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER;
-        hash = (53 * hash) + getServerName().hashCode();
-      }
-      if (hasTimestamp()) {
-        hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER;
-        hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
-            getTimestamp());
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * <pre>
-     **
-     * ServerTimestamp keeps last WAL roll time per Region Server
-     * </pre>
-     *
-     * Protobuf type {@code hbase.pb.ServerTimestamp}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:hbase.pb.ServerTimestamp)
-        org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_ServerTimestamp_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_ServerTimestamp_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.class, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-          getServerNameFieldBuilder();
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        if (serverNameBuilder_ == null) {
-          serverName_ = null;
-        } else {
-          serverNameBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        timestamp_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_ServerTimestamp_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp build() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp buildPartial() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp result = new org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        if (serverNameBuilder_ == null) {
-          result.serverName_ = serverName_;
-        } else {
-          result.serverName_ = serverNameBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.timestamp_ = timestamp_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp other) {
-        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.getDefaultInstance()) return this;
-        if (other.hasServerName()) {
-          mergeServerName(other.getServerName());
-        }
-        if (other.hasTimestamp()) {
-          setTimestamp(other.getTimestamp());
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (hasServerName()) {
-          if (!getServerName().isInitialized()) {
-            return false;
-          }
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_ = null;
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_;
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      public boolean hasServerName() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName() {
-        if (serverNameBuilder_ == null) {
-          return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_;
-        } else {
-          return serverNameBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      public Builder setServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) {
-        if (serverNameBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          serverName_ = value;
-          onChanged();
-        } else {
-          serverNameBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      public Builder setServerName(
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) {
-        if (serverNameBuilder_ == null) {
-          serverName_ = builderForValue.build();
-          onChanged();
-        } else {
-          serverNameBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      public Builder mergeServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) {
-        if (serverNameBuilder_ == null) {
-          if (((bitField0_ & 0x00000001) == 0x00000001) &&
-              serverName_ != null &&
-              serverName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) {
-            serverName_ =
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.newBuilder(serverName_).mergeFrom(value).buildPartial();
-          } else {
-            serverName_ = value;
-          }
-          onChanged();
-        } else {
-          serverNameBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      public Builder clearServerName() {
-        if (serverNameBuilder_ == null) {
-          serverName_ = null;
-          onChanged();
-        } else {
-          serverNameBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder() {
-        bitField0_ |= 0x00000001;
-        onChanged();
-        return getServerNameFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() {
-        if (serverNameBuilder_ != null) {
-          return serverNameBuilder_.getMessageOrBuilder();
-        } else {
-          return serverName_ == null ?
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_;
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.ServerName server_name = 1;</code>
-       */
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> 
-          getServerNameFieldBuilder() {
-        if (serverNameBuilder_ == null) {
-          serverNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>(
-                  getServerName(),
-                  getParentForChildren(),
-                  isClean());
-          serverName_ = null;
-        }
-        return serverNameBuilder_;
-      }
-
-      private long timestamp_ ;
-      /**
-       * <code>optional uint64 timestamp = 2;</code>
-       */
-      public boolean hasTimestamp() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional uint64 timestamp = 2;</code>
-       */
-      public long getTimestamp() {
-        return timestamp_;
-      }
-      /**
-       * <code>optional uint64 timestamp = 2;</code>
-       */
-      public Builder setTimestamp(long value) {
-        bitField0_ |= 0x00000002;
-        timestamp_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional uint64 timestamp = 2;</code>
-       */
-      public Builder clearTimestamp() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        timestamp_ = 0L;
-        onChanged();
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.ServerTimestamp)
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.ServerTimestamp)
-    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ServerTimestamp>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ServerTimestamp>() {
-      public ServerTimestamp parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new ServerTimestamp(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ServerTimestamp> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ServerTimestamp> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface TableServerTimestampOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:hbase.pb.TableServerTimestamp)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    boolean hasTableName();
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName();
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
-
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp> 
-        getServerTimestampList();
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp getServerTimestamp(int index);
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    int getServerTimestampCount();
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder> 
-        getServerTimestampOrBuilderList();
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder getServerTimestampOrBuilder(
-        int index);
-  }
-  /**
-   * <pre>
-   **
-   *  TableServerTimestamp keeps last WAL roll time per Region Server &amp; Table
-   *  Each table have different last WAL roll time stamps across cluster, on every RS
-   * </pre>
-   *
-   * Protobuf type {@code hbase.pb.TableServerTimestamp}
-   */
-  public  static final class TableServerTimestamp extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:hbase.pb.TableServerTimestamp)
-      TableServerTimestampOrBuilder {
-    // Use TableServerTimestamp.newBuilder() to construct.
-    private TableServerTimestamp(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private TableServerTimestamp() {
-      serverTimestamp_ = java.util.Collections.emptyList();
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private TableServerTimestamp(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = tableName_.toBuilder();
-              }
-              tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(tableName_);
-                tableName_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000001;
-              break;
-            }
-            case 18: {
-              if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-                serverTimestamp_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp>();
-                mutable_bitField0_ |= 0x00000002;
-              }
-              serverTimestamp_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.PARSER, extensionRegistry));
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-          serverTimestamp_ = java.util.Collections.unmodifiableList(serverTimestamp_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_TableServerTimestamp_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_TableServerTimestamp_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp.class, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int TABLE_NAME_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    public boolean hasTableName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
-      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
-    }
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
-      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
-    }
-
-    public static final int SERVER_TIMESTAMP_FIELD_NUMBER = 2;
-    private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp> serverTimestamp_;
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp> getServerTimestampList() {
-      return serverTimestamp_;
-    }
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder> 
-        getServerTimestampOrBuilderList() {
-      return serverTimestamp_;
-    }
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    public int getServerTimestampCount() {
-      return serverTimestamp_.size();
-    }
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp getServerTimestamp(int index) {
-      return serverTimestamp_.get(index);
-    }
-    /**
-     * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder getServerTimestampOrBuilder(
-        int index) {
-      return serverTimestamp_.get(index);
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (hasTableName()) {
-        if (!getTableName().isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      for (int i = 0; i < getServerTimestampCount(); i++) {
-        if (!getServerTimestamp(i).isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, getTableName());
-      }
-      for (int i = 0; i < serverTimestamp_.size(); i++) {
-        output.writeMessage(2, serverTimestamp_.get(i));
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, getTableName());
-      }
-      for (int i = 0; i < serverTimestamp_.size(); i++) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(2, serverTimestamp_.get(i));
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp other = (org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp) obj;
-
-      boolean result = true;
-      result = result && (hasTableName() == other.hasTableName());
-      if (hasTableName()) {
-        result = result && getTableName()
-            .equals(other.getTableName());
-      }
-      result = result && getServerTimestampList()
-          .equals(other.getServerTimestampList());
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasTableName()) {
-        hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
-        hash = (53 * hash) + getTableName().hashCode();
-      }
-      if (getServerTimestampCount() > 0) {
-        hash = (37 * hash) + SERVER_TIMESTAMP_FIELD_NUMBER;
-        hash = (53 * hash) + getServerTimestampList().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * <pre>
-     **
-     *  TableServerTimestamp keeps last WAL roll time per Region Server &amp; Table
-     *  Each table have different last WAL roll time stamps across cluster, on every RS
-     * </pre>
-     *
-     * Protobuf type {@code hbase.pb.TableServerTimestamp}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:hbase.pb.TableServerTimestamp)
-        org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_TableServerTimestamp_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_TableServerTimestamp_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp.class, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-          getTableNameFieldBuilder();
-          getServerTimestampFieldBuilder();
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        if (tableNameBuilder_ == null) {
-          tableName_ = null;
-        } else {
-          tableNameBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        if (serverTimestampBuilder_ == null) {
-          serverTimestamp_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000002);
-        } else {
-          serverTimestampBuilder_.clear();
-        }
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.internal_static_hbase_pb_TableServerTimestamp_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp build() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp buildPartial() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp result = new org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        if (tableNameBuilder_ == null) {
-          result.tableName_ = tableName_;
-        } else {
-          result.tableName_ = tableNameBuilder_.build();
-        }
-        if (serverTimestampBuilder_ == null) {
-          if (((bitField0_ & 0x00000002) == 0x00000002)) {
-            serverTimestamp_ = java.util.Collections.unmodifiableList(serverTimestamp_);
-            bitField0_ = (bitField0_ & ~0x00000002);
-          }
-          result.serverTimestamp_ = serverTimestamp_;
-        } else {
-          result.serverTimestamp_ = serverTimestampBuilder_.build();
-        }
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp other) {
-        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp.getDefaultInstance()) return this;
-        if (other.hasTableName()) {
-          mergeTableName(other.getTableName());
-        }
-        if (serverTimestampBuilder_ == null) {
-          if (!other.serverTimestamp_.isEmpty()) {
-            if (serverTimestamp_.isEmpty()) {
-              serverTimestamp_ = other.serverTimestamp_;
-              bitField0_ = (bitField0_ & ~0x00000002);
-            } else {
-              ensureServerTimestampIsMutable();
-              serverTimestamp_.addAll(other.serverTimestamp_);
-            }
-            onChanged();
-          }
-        } else {
-          if (!other.serverTimestamp_.isEmpty()) {
-            if (serverTimestampBuilder_.isEmpty()) {
-              serverTimestampBuilder_.dispose();
-              serverTimestampBuilder_ = null;
-              serverTimestamp_ = other.serverTimestamp_;
-              bitField0_ = (bitField0_ & ~0x00000002);
-              serverTimestampBuilder_ = 
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                   getServerTimestampFieldBuilder() : null;
-            } else {
-              serverTimestampBuilder_.addAllMessages(other.serverTimestamp_);
-            }
-          }
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (hasTableName()) {
-          if (!getTableName().isInitialized()) {
-            return false;
-          }
-        }
-        for (int i = 0; i < getServerTimestampCount(); i++) {
-          if (!getServerTimestamp(i).isInitialized()) {
-            return false;
-          }
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_ = null;
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      public boolean hasTableName() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
-        if (tableNameBuilder_ == null) {
-          return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
-        } else {
-          return tableNameBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
-        if (tableNameBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          tableName_ = value;
-          onChanged();
-        } else {
-          tableNameBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      public Builder setTableName(
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
-        if (tableNameBuilder_ == null) {
-          tableName_ = builderForValue.build();
-          onChanged();
-        } else {
-          tableNameBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
-        if (tableNameBuilder_ == null) {
-          if (((bitField0_ & 0x00000001) == 0x00000001) &&
-              tableName_ != null &&
-              tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
-            tableName_ =
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
-          } else {
-            tableName_ = value;
-          }
-          onChanged();
-        } else {
-          tableNameBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      public Builder clearTableName() {
-        if (tableNameBuilder_ == null) {
-          tableName_ = null;
-          onChanged();
-        } else {
-          tableNameBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
-        bitField0_ |= 0x00000001;
-        onChanged();
-        return getTableNameFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
-        if (tableNameBuilder_ != null) {
-          return tableNameBuilder_.getMessageOrBuilder();
-        } else {
-          return tableName_ == null ?
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.TableName table_name = 1;</code>
-       */
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> 
-          getTableNameFieldBuilder() {
-        if (tableNameBuilder_ == null) {
-          tableNameBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
-                  getTableName(),
-                  getParentForChildren(),
-                  isClean());
-          tableName_ = null;
-        }
-        return tableNameBuilder_;
-      }
-
-      private java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp> serverTimestamp_ =
-        java.util.Collections.emptyList();
-      private void ensureServerTimestampIsMutable() {
-        if (!((bitField0_ & 0x00000002) == 0x00000002)) {
-          serverTimestamp_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp>(serverTimestamp_);
-          bitField0_ |= 0x00000002;
-         }
-      }
-
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder> serverTimestampBuilder_;
-
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp> getServerTimestampList() {
-        if (serverTimestampBuilder_ == null) {
-          return java.util.Collections.unmodifiableList(serverTimestamp_);
-        } else {
-          return serverTimestampBuilder_.getMessageList();
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public int getServerTimestampCount() {
-        if (serverTimestampBuilder_ == null) {
-          return serverTimestamp_.size();
-        } else {
-          return serverTimestampBuilder_.getCount();
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp getServerTimestamp(int index) {
-        if (serverTimestampBuilder_ == null) {
-          return serverTimestamp_.get(index);
-        } else {
-          return serverTimestampBuilder_.getMessage(index);
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder setServerTimestamp(
-          int index, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp value) {
-        if (serverTimestampBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureServerTimestampIsMutable();
-          serverTimestamp_.set(index, value);
-          onChanged();
-        } else {
-          serverTimestampBuilder_.setMessage(index, value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder setServerTimestamp(
-          int index, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder builderForValue) {
-        if (serverTimestampBuilder_ == null) {
-          ensureServerTimestampIsMutable();
-          serverTimestamp_.set(index, builderForValue.build());
-          onChanged();
-        } else {
-          serverTimestampBuilder_.setMessage(index, builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder addServerTimestamp(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp value) {
-        if (serverTimestampBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureServerTimestampIsMutable();
-          serverTimestamp_.add(value);
-          onChanged();
-        } else {
-          serverTimestampBuilder_.addMessage(value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder addServerTimestamp(
-          int index, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp value) {
-        if (serverTimestampBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureServerTimestampIsMutable();
-          serverTimestamp_.add(index, value);
-          onChanged();
-        } else {
-          serverTimestampBuilder_.addMessage(index, value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder addServerTimestamp(
-          org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder builderForValue) {
-        if (serverTimestampBuilder_ == null) {
-          ensureServerTimestampIsMutable();
-          serverTimestamp_.add(builderForValue.build());
-          onChanged();
-        } else {
-          serverTimestampBuilder_.addMessage(builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder addServerTimestamp(
-          int index, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder builderForValue) {
-        if (serverTimestampBuilder_ == null) {
-          ensureServerTimestampIsMutable();
-          serverTimestamp_.add(index, builderForValue.build());
-          onChanged();
-        } else {
-          serverTimestampBuilder_.addMessage(index, builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder addAllServerTimestamp(
-          java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp> values) {
-        if (serverTimestampBuilder_ == null) {
-          ensureServerTimestampIsMutable();
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-              values, serverTimestamp_);
-          onChanged();
-        } else {
-          serverTimestampBuilder_.addAllMessages(values);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder clearServerTimestamp() {
-        if (serverTimestampBuilder_ == null) {
-          serverTimestamp_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000002);
-          onChanged();
-        } else {
-          serverTimestampBuilder_.clear();
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public Builder removeServerTimestamp(int index) {
-        if (serverTimestampBuilder_ == null) {
-          ensureServerTimestampIsMutable();
-          serverTimestamp_.remove(index);
-          onChanged();
-        } else {
-          serverTimestampBuilder_.remove(index);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder getServerTimestampBuilder(
-          int index) {
-        return getServerTimestampFieldBuilder().getBuilder(index);
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder getServerTimestampOrBuilder(
-          int index) {
-        if (serverTimestampBuilder_ == null) {
-          return serverTimestamp_.get(index);  } else {
-          return serverTimestampBuilder_.getMessageOrBuilder(index);
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder> 
-           getServerTimestampOrBuilderList() {
-        if (serverTimestampBuilder_ != null) {
-          return serverTimestampBuilder_.getMessageOrBuilderList();
-        } else {
-          return java.util.Collections.unmodifiableList(serverTimestamp_);
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder addServerTimestampBuilder() {
-        return getServerTimestampFieldBuilder().addBuilder(
-            org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.getDefaultInstance());
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder addServerTimestampBuilder(
-          int index) {
-        return getServerTimestampFieldBuilder().addBuilder(
-            index, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.getDefaultInstance());
-      }
-      /**
-       * <code>repeated .hbase.pb.ServerTimestamp server_timestamp = 2;</code>
-       */
-      public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder> 
-           getServerTimestampBuilderList() {
-        return getServerTimestampFieldBuilder().getBuilderList();
-      }
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder> 
-          getServerTimestampFieldBuilder() {
-        if (serverTimestampBuilder_ == null) {
-          serverTimestampBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestamp.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.ServerTimestampOrBuilder>(
-                  serverTimestamp_,
-                  ((bitField0_ & 0x00000002) == 0x00000002),
-                  getParentForChildren(),
-                  isClean());
-          serverTimestamp_ = null;
-        }
-        return serverTimestampBuilder_;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.TableServerTimestamp)
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.TableServerTimestamp)
-    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableServerTimestamp>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<TableServerTimestamp>() {
-      public TableServerTimestamp parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new TableServerTimestamp(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableServerTimestamp> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TableServerTimestamp> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface BackupImageOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:hbase.pb.BackupImage)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>optional string backup_id = 1;</code>
-     */
-    boolean hasBackupId();
-    /**
-     * <code>optional string backup_id = 1;</code>
-     */
-    java.lang.String getBackupId();
-    /**
-     * <code>optional string backup_id = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getBackupIdBytes();
-
-    /**
-     * <code>optional .hbase.pb.BackupType backup_type = 2;</code>
-     */
-    boolean hasBackupType();
-    /**
-     * <code>optional .hbase.pb.BackupType backup_type = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupType getBackupType();
-
-    /**
-     * <code>optional string backup_root_dir = 3;</code>
-     */
-    boolean hasBackupRootDir();
-    /**
-     * <code>optional string backup_root_dir = 3;</code>
-     */
-    java.lang.String getBackupRootDir();
-    /**
-     * <code>optional string backup_root_dir = 3;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getBackupRootDirBytes();
-
-    /**
-     * <code>repeated .hbase.pb.TableName table_list = 4;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName> 
-        getTableListList();
-    /**
-     * <code>repeated .hbase.pb.TableName table_list = 4;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableList(int index);
-    /**
-     * <code>repeated .hbase.pb.TableName table_list = 4;</code>
-     */
-    int getTableListCount();
-    /**
-     * <code>repeated .hbase.pb.TableName table_list = 4;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> 
-        getTableListOrBuilderList();
-    /**
-     * <code>repeated .hbase.pb.TableName table_list = 4;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableListOrBuilder(
-        int index);
-
-    /**
-     * <code>optional uint64 start_ts = 5;</code>
-     */
-    boolean hasStartTs();
-    /**
-     * <code>optional uint64 start_ts = 5;</code>
-     */
-    long getStartTs();
-
-    /**
-     * <code>optional uint64 complete_ts = 6;</code>
-     */
-    boolean hasCompleteTs();
-    /**
-     * <code>optional uint64 complete_ts = 6;</code>
-     */
-    long getCompleteTs();
-
-    /**
-     * <code>repeated .hbase.pb.BackupImage ancestors = 7;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupImage> 
-        getAncestorsList();
-    /**
-     * <code>repeated .hbase.pb.BackupImage ancestors = 7;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupImage getAncestors(int index);
-    /**
-     * <code>repeated .hbase.pb.BackupImage ancestors = 7;</code>
-     */
-    int getAncestorsCount();
-    /**
-     * <code>repeated .hbase.pb.BackupImage ancestors = 7;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupImageOrBuilder> 
-        getAncestorsOrBuilderList();
-    /**
-     * <code>repeated .hbase.pb.BackupImage ancestors = 7;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupImageOrBuilder getAncestorsOrBuilder(
-        int index);
-
-    /**
-     * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 8;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp> 
-        getTstMapList();
-    /**
-     * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 8;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp getTstMap(int index);
-    /**
-     * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 8;</code>
-     */
-    int getTstMapCount();
-    /**
-     * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 8;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder> 
-        getTstMapOrBuilderList();
-    /**
-     * <code>repeated .hbase.pb.TableServerTimestamp tst_map = 8;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestampOrBuilder getTstMapOrBuilder(
-        int index);
-  }
-  /**
-   * <pre>
-   **
-   * Structure keeps relevant info for backup restore session
-   * </pre>
-   *
-   * Protobuf type {@code hbase.pb.BackupImage}
-   */
-  public  static final class BackupImage extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:hbase.pb.BackupImage)
-      BackupImageOrBuilder {
-    // Use BackupImage.newBuilder() to construct.
-    private BackupImage(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private BackupImage() {
-      backupId_ = "";
-      backupType_ = 0;
-      backupRootDir_ = "";
-      tableList_ = java.util.Collections.emptyList();
-      startTs_ = 0L;
-      completeTs_ = 0L;
-      ancestors_ = java.util.Collections.emptyList();
-      tstMap_ = java.util.Collections.emptyList();
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private BackupImage(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000001;
-              backupId_ = bs;
-              break;
-            }
-            case 16: {
-              int rawValue = input.readEnum();
-              org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupType value = org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupType.valueOf(rawValue);
-              if (value == null) {
-                unknownFields.mergeVarintField(2, rawValue);
-              } else {
-                bitField0_ |= 0x00000002;
-                backupType_ = rawValue;
-              }
-              break;
-            }
-            case 26: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000004;
-              backupRootDir_ = bs;
-              break;
-            }
-            case 34: {
-              if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-                tableList_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName>();
-                mutable_bitField0_ |= 0x00000008;
-              }
-              tableList_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry));
-              break;
-            }
-            case 40: {
-              bitField0_ |= 0x00000008;
-              startTs_ = input.readUInt64();
-              break;
-            }
-            case 48: {
-              bitField0_ |= 0x00000010;
-              completeTs_ = input.readUInt64();
-              break;
-            }
-            case 58: {
-              if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
-                ancestors_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupImage>();
-                mutable_bitField0_ |= 0x00000040;
-              }
-              ancestors_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupImage.PARSER, extensionRegistry));
-              break;
-            }
-            case 66: {
-              if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) {
-                tstMap_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp>();
-                mutable_bitField0_ |= 0x00000080;
-              }
-              tstMap_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp.PARSER, extensionRegistry));
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-          tableList_ = java.util.Collections.unmodifiableList(tableList_);
-        }
-        if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
-          ancestors_ = java.util.Collections.unmodifiableList(ancestors_);
-        }
-        if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) {
-          tstMap_ = java.util.Collections.unmodifiableList(tstMap_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.D

<TRUNCATED>

[14/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormat.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormat.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormat.java
deleted file mode 100644
index 491089b..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormat.java
+++ /dev/null
@@ -1,2108 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-
-import java.io.IOException;
-import java.math.BigInteger;
-import java.nio.CharBuffer;
-import java.util.ArrayList;
-import java.util.List;
-import java.util.Locale;
-import java.util.Map;
-import java.util.logging.Logger;
-import java.util.regex.Matcher;
-import java.util.regex.Pattern;
-
-/**
- * Provide text parsing and formatting support for proto2 instances.
- * The implementation largely follows google/protobuf/text_format.cc.
- *
- * @author wenboz@google.com Wenbo Zhu
- * @author kenton@google.com Kenton Varda
- */
-public final class TextFormat {
-  private TextFormat() {}
-
-  private static final Logger logger =
-      Logger.getLogger(TextFormat.class.getName());
-
-  private static final Printer DEFAULT_PRINTER = new Printer();
-  private static final Printer SINGLE_LINE_PRINTER =
-      (new Printer()).setSingleLineMode(true);
-  private static final Printer UNICODE_PRINTER =
-      (new Printer()).setEscapeNonAscii(false);
-
-  /**
-   * Outputs a textual representation of the Protocol Message supplied into
-   * the parameter output. (This representation is the new version of the
-   * classic "ProtocolPrinter" output from the original Protocol Buffer system)
-   */
-  public static void print(
-      final MessageOrBuilder message, final Appendable output)
-      throws IOException {
-    DEFAULT_PRINTER.print(message, new TextGenerator(output));
-  }
-
-  /** Outputs a textual representation of {@code fields} to {@code output}. */
-  public static void print(final UnknownFieldSet fields,
-                           final Appendable output)
-                           throws IOException {
-    DEFAULT_PRINTER.printUnknownFields(fields, new TextGenerator(output));
-  }
-
-  /**
-   * Same as {@code print()}, except that non-ASCII characters are not
-   * escaped.
-   */
-  public static void printUnicode(
-      final MessageOrBuilder message, final Appendable output)
-      throws IOException {
-    UNICODE_PRINTER.print(message, new TextGenerator(output));
-  }
-
-  /**
-   * Same as {@code print()}, except that non-ASCII characters are not
-   * escaped.
-   */
-  public static void printUnicode(final UnknownFieldSet fields,
-                                  final Appendable output)
-                                  throws IOException {
-    UNICODE_PRINTER.printUnknownFields(fields, new TextGenerator(output));
-  }
-
-  /**
-   * Generates a human readable form of this message, useful for debugging and
-   * other purposes, with no newline characters.
-   */
-  public static String shortDebugString(final MessageOrBuilder message) {
-    try {
-      final StringBuilder sb = new StringBuilder();
-      SINGLE_LINE_PRINTER.print(message, new TextGenerator(sb));
-      // Single line mode currently might have an extra space at the end.
-      return sb.toString().trim();
-    } catch (IOException e) {
-      throw new IllegalStateException(e);
-    }
-  }
-
-  /**
-   * Generates a human readable form of the field, useful for debugging
-   * and other purposes, with no newline characters.
-   */
-  public static String shortDebugString(final FieldDescriptor field,
-                                        final Object value) {
-    try {
-      final StringBuilder sb = new StringBuilder();
-      SINGLE_LINE_PRINTER.printField(field, value, new TextGenerator(sb));
-      return sb.toString().trim();
-    } catch (IOException e) {
-        throw new IllegalStateException(e);
-    }
-  }
-
-  /**
-   * Generates a human readable form of the unknown fields, useful for debugging
-   * and other purposes, with no newline characters.
-   */
-  public static String shortDebugString(final UnknownFieldSet fields) {
-    try {
-      final StringBuilder sb = new StringBuilder();
-      SINGLE_LINE_PRINTER.printUnknownFields(fields, new TextGenerator(sb));
-      // Single line mode currently might have an extra space at the end.
-      return sb.toString().trim();
-    } catch (IOException e) {
-      throw new IllegalStateException(e);
-    }
-  }
-
-  /**
-   * Like {@code print()}, but writes directly to a {@code String} and
-   * returns it.
-   */
-  public static String printToString(final MessageOrBuilder message) {
-    try {
-      final StringBuilder text = new StringBuilder();
-      print(message, text);
-      return text.toString();
-    } catch (IOException e) {
-      throw new IllegalStateException(e);
-    }
-  }
-
-  /**
-   * Like {@code print()}, but writes directly to a {@code String} and
-   * returns it.
-   */
-  public static String printToString(final UnknownFieldSet fields) {
-    try {
-      final StringBuilder text = new StringBuilder();
-      print(fields, text);
-      return text.toString();
-    } catch (IOException e) {
-      throw new IllegalStateException(e);
-    }
-  }
-
-  /**
-   * Same as {@code printToString()}, except that non-ASCII characters
-   * in string type fields are not escaped in backslash+octals.
-   */
-  public static String printToUnicodeString(final MessageOrBuilder message) {
-    try {
-      final StringBuilder text = new StringBuilder();
-      UNICODE_PRINTER.print(message, new TextGenerator(text));
-      return text.toString();
-    } catch (IOException e) {
-      throw new IllegalStateException(e);
-    }
-  }
-
-  /**
-   * Same as {@code printToString()}, except that non-ASCII characters
-   * in string type fields are not escaped in backslash+octals.
-   */
-  public static String printToUnicodeString(final UnknownFieldSet fields) {
-    try {
-      final StringBuilder text = new StringBuilder();
-      UNICODE_PRINTER.printUnknownFields(fields, new TextGenerator(text));
-      return text.toString();
-    } catch (IOException e) {
-      throw new IllegalStateException(e);
-    }
-  }
-
-  public static void printField(final FieldDescriptor field,
-                                final Object value,
-                                final Appendable output)
-                                throws IOException {
-    DEFAULT_PRINTER.printField(field, value, new TextGenerator(output));
-  }
-
-  public static String printFieldToString(final FieldDescriptor field,
-                                          final Object value) {
-    try {
-      final StringBuilder text = new StringBuilder();
-      printField(field, value, text);
-      return text.toString();
-    } catch (IOException e) {
-      throw new IllegalStateException(e);
-    }
-  }
-
-  /**
-   * Outputs a textual representation of the value of given field value.
-   *
-   * @param field the descriptor of the field
-   * @param value the value of the field
-   * @param output the output to which to append the formatted value
-   * @throws ClassCastException if the value is not appropriate for the
-   *     given field descriptor
-   * @throws IOException if there is an exception writing to the output
-   */
-  public static void printFieldValue(final FieldDescriptor field,
-                                     final Object value,
-                                     final Appendable output)
-                                     throws IOException {
-    DEFAULT_PRINTER.printFieldValue(field, value, new TextGenerator(output));
-  }
-
-  /**
-   * Outputs a textual representation of the value of an unknown field.
-   *
-   * @param tag the field's tag number
-   * @param value the value of the field
-   * @param output the output to which to append the formatted value
-   * @throws ClassCastException if the value is not appropriate for the
-   *     given field descriptor
-   * @throws IOException if there is an exception writing to the output
-   */
-  public static void printUnknownFieldValue(final int tag,
-                                            final Object value,
-                                            final Appendable output)
-                                            throws IOException {
-    printUnknownFieldValue(tag, value, new TextGenerator(output));
-  }
-
-  private static void printUnknownFieldValue(final int tag,
-                                             final Object value,
-                                             final TextGenerator generator)
-                                             throws IOException {
-    switch (WireFormat.getTagWireType(tag)) {
-      case WireFormat.WIRETYPE_VARINT:
-        generator.print(unsignedToString((Long) value));
-        break;
-      case WireFormat.WIRETYPE_FIXED32:
-        generator.print(
-            String.format((Locale) null, "0x%08x", (Integer) value));
-        break;
-      case WireFormat.WIRETYPE_FIXED64:
-        generator.print(String.format((Locale) null, "0x%016x", (Long) value));
-        break;
-      case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-        generator.print("\"");
-        generator.print(escapeBytes((ByteString) value));
-        generator.print("\"");
-        break;
-      case WireFormat.WIRETYPE_START_GROUP:
-        DEFAULT_PRINTER.printUnknownFields((UnknownFieldSet) value, generator);
-        break;
-      default:
-        throw new IllegalArgumentException("Bad tag: " + tag);
-    }
-  }
-
-  /** Helper class for converting protobufs to text. */
-  private static final class Printer {
-    /** Whether to omit newlines from the output. */
-    boolean singleLineMode = false;
-
-    /** Whether to escape non ASCII characters with backslash and octal. */
-    boolean escapeNonAscii = true;
-
-    private Printer() {}
-
-    /** Setter of singleLineMode */
-    private Printer setSingleLineMode(boolean singleLineMode) {
-      this.singleLineMode = singleLineMode;
-      return this;
-    }
-
-    /** Setter of escapeNonAscii */
-    private Printer setEscapeNonAscii(boolean escapeNonAscii) {
-      this.escapeNonAscii = escapeNonAscii;
-      return this;
-    }
-
-    private void print(
-        final MessageOrBuilder message, final TextGenerator generator)
-        throws IOException {
-      for (Map.Entry<FieldDescriptor, Object> field
-          : message.getAllFields().entrySet()) {
-        printField(field.getKey(), field.getValue(), generator);
-      }
-      printUnknownFields(message.getUnknownFields(), generator);
-    }
-
-    private void printField(final FieldDescriptor field, final Object value,
-        final TextGenerator generator) throws IOException {
-      if (field.isRepeated()) {
-        // Repeated field.  Print each element.
-        for (Object element : (List<?>) value) {
-          printSingleField(field, element, generator);
-        }
-      } else {
-        printSingleField(field, value, generator);
-      }
-    }
-
-    private void printSingleField(final FieldDescriptor field,
-                                  final Object value,
-                                  final TextGenerator generator)
-                                  throws IOException {
-      if (field.isExtension()) {
-        generator.print("[");
-        // We special-case MessageSet elements for compatibility with proto1.
-        if (field.getContainingType().getOptions().getMessageSetWireFormat()
-            && (field.getType() == FieldDescriptor.Type.MESSAGE)
-            && (field.isOptional())
-            // object equality
-            && (field.getExtensionScope() == field.getMessageType())) {
-          generator.print(field.getMessageType().getFullName());
-        } else {
-          generator.print(field.getFullName());
-        }
-        generator.print("]");
-      } else {
-        if (field.getType() == FieldDescriptor.Type.GROUP) {
-          // Groups must be serialized with their original capitalization.
-          generator.print(field.getMessageType().getName());
-        } else {
-          generator.print(field.getName());
-        }
-      }
-
-      if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-        if (singleLineMode) {
-          generator.print(" { ");
-        } else {
-          generator.print(" {\n");
-          generator.indent();
-        }
-      } else {
-        generator.print(": ");
-      }
-
-      printFieldValue(field, value, generator);
-
-      if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-        if (singleLineMode) {
-          generator.print("} ");
-        } else {
-          generator.outdent();
-          generator.print("}\n");
-        }
-      } else {
-        if (singleLineMode) {
-          generator.print(" ");
-        } else {
-          generator.print("\n");
-        }
-      }
-    }
-
-    private void printFieldValue(final FieldDescriptor field,
-                                 final Object value,
-                                 final TextGenerator generator)
-                                 throws IOException {
-      switch (field.getType()) {
-        case INT32:
-        case SINT32:
-        case SFIXED32:
-          generator.print(((Integer) value).toString());
-          break;
-
-        case INT64:
-        case SINT64:
-        case SFIXED64:
-          generator.print(((Long) value).toString());
-          break;
-
-        case BOOL:
-          generator.print(((Boolean) value).toString());
-          break;
-
-        case FLOAT:
-          generator.print(((Float) value).toString());
-          break;
-
-        case DOUBLE:
-          generator.print(((Double) value).toString());
-          break;
-
-        case UINT32:
-        case FIXED32:
-          generator.print(unsignedToString((Integer) value));
-          break;
-
-        case UINT64:
-        case FIXED64:
-          generator.print(unsignedToString((Long) value));
-          break;
-
-        case STRING:
-          generator.print("\"");
-          generator.print(escapeNonAscii
-              ? TextFormatEscaper.escapeText((String) value)
-              : escapeDoubleQuotesAndBackslashes((String) value)
-                  .replace("\n", "\\n"));
-          generator.print("\"");
-          break;
-
-        case BYTES:
-          generator.print("\"");
-          if (value instanceof ByteString) {
-            generator.print(escapeBytes((ByteString) value));
-          } else {
-            generator.print(escapeBytes((byte[]) value));
-          }
-          generator.print("\"");
-          break;
-
-        case ENUM:
-          generator.print(((EnumValueDescriptor) value).getName());
-          break;
-
-        case MESSAGE:
-        case GROUP:
-          print((Message) value, generator);
-          break;
-      }
-    }
-
-    private void printUnknownFields(final UnknownFieldSet unknownFields,
-                                    final TextGenerator generator)
-                                    throws IOException {
-      for (Map.Entry<Integer, UnknownFieldSet.Field> entry :
-               unknownFields.asMap().entrySet()) {
-        final int number = entry.getKey();
-        final UnknownFieldSet.Field field = entry.getValue();
-        printUnknownField(number, WireFormat.WIRETYPE_VARINT,
-            field.getVarintList(), generator);
-        printUnknownField(number, WireFormat.WIRETYPE_FIXED32,
-            field.getFixed32List(), generator);
-        printUnknownField(number, WireFormat.WIRETYPE_FIXED64,
-            field.getFixed64List(), generator);
-        printUnknownField(number, WireFormat.WIRETYPE_LENGTH_DELIMITED,
-            field.getLengthDelimitedList(), generator);
-        for (final UnknownFieldSet value : field.getGroupList()) {
-          generator.print(entry.getKey().toString());
-          if (singleLineMode) {
-            generator.print(" { ");
-          } else {
-            generator.print(" {\n");
-            generator.indent();
-          }
-          printUnknownFields(value, generator);
-          if (singleLineMode) {
-            generator.print("} ");
-          } else {
-            generator.outdent();
-            generator.print("}\n");
-          }
-        }
-      }
-    }
-
-    private void printUnknownField(final int number,
-                                   final int wireType,
-                                   final List<?> values,
-                                   final TextGenerator generator)
-                                   throws IOException {
-      for (final Object value : values) {
-        generator.print(String.valueOf(number));
-        generator.print(": ");
-        printUnknownFieldValue(wireType, value, generator);
-        generator.print(singleLineMode ? " " : "\n");
-      }
-    }
-  }
-
-  /** Convert an unsigned 32-bit integer to a string. */
-  public static String unsignedToString(final int value) {
-    if (value >= 0) {
-      return Integer.toString(value);
-    } else {
-      return Long.toString(value & 0x00000000FFFFFFFFL);
-    }
-  }
-
-  /** Convert an unsigned 64-bit integer to a string. */
-  public static String unsignedToString(final long value) {
-    if (value >= 0) {
-      return Long.toString(value);
-    } else {
-      // Pull off the most-significant bit so that BigInteger doesn't think
-      // the number is negative, then set it again using setBit().
-      return BigInteger.valueOf(value & 0x7FFFFFFFFFFFFFFFL)
-                       .setBit(63).toString();
-    }
-  }
-
-  /**
-   * An inner class for writing text to the output stream.
-   */
-  private static final class TextGenerator {
-    private final Appendable output;
-    private final StringBuilder indent = new StringBuilder();
-    private boolean atStartOfLine = true;
-
-    private TextGenerator(final Appendable output) {
-      this.output = output;
-    }
-
-    /**
-     * Indent text by two spaces.  After calling Indent(), two spaces will be
-     * inserted at the beginning of each line of text.  Indent() may be called
-     * multiple times to produce deeper indents.
-     */
-    public void indent() {
-      indent.append("  ");
-    }
-
-    /**
-     * Reduces the current indent level by two spaces, or crashes if the indent
-     * level is zero.
-     */
-    public void outdent() {
-      final int length = indent.length();
-      if (length == 0) {
-        throw new IllegalArgumentException(
-            " Outdent() without matching Indent().");
-      }
-      indent.delete(length - 2, length);
-    }
-
-    /**
-     * Print text to the output stream.
-     */
-    public void print(final CharSequence text) throws IOException {
-      final int size = text.length();
-      int pos = 0;
-
-      for (int i = 0; i < size; i++) {
-        if (text.charAt(i) == '\n') {
-          write(text.subSequence(pos, i + 1));
-          pos = i + 1;
-          atStartOfLine = true;
-        }
-      }
-      write(text.subSequence(pos, size));
-    }
-
-    private void write(final CharSequence data) throws IOException {
-      if (data.length() == 0) {
-        return;
-      }
-      if (atStartOfLine) {
-        atStartOfLine = false;
-        output.append(indent);
-      }
-      output.append(data);
-    }
-  }
-
-  // =================================================================
-  // Parsing
-
-  /**
-   * Represents a stream of tokens parsed from a {@code String}.
-   *
-   * <p>The Java standard library provides many classes that you might think
-   * would be useful for implementing this, but aren't.  For example:
-   *
-   * <ul>
-   * <li>{@code java.io.StreamTokenizer}:  This almost does what we want -- or,
-   *   at least, something that would get us close to what we want -- except
-   *   for one fatal flaw:  It automatically un-escapes strings using Java
-   *   escape sequences, which do not include all the escape sequences we
-   *   need to support (e.g. '\x').
-   * <li>{@code java.util.Scanner}:  This seems like a great way at least to
-   *   parse regular expressions out of a stream (so we wouldn't have to load
-   *   the entire input into a single string before parsing).  Sadly,
-   *   {@code Scanner} requires that tokens be delimited with some delimiter.
-   *   Thus, although the text "foo:" should parse to two tokens ("foo" and
-   *   ":"), {@code Scanner} would recognize it only as a single token.
-   *   Furthermore, {@code Scanner} provides no way to inspect the contents
-   *   of delimiters, making it impossible to keep track of line and column
-   *   numbers.
-   * </ul>
-   *
-   * <p>Luckily, Java's regular expression support does manage to be useful to
-   * us.  (Barely:  We need {@code Matcher.usePattern()}, which is new in
-   * Java 1.5.)  So, we can use that, at least.  Unfortunately, this implies
-   * that we need to have the entire input in one contiguous string.
-   */
-  private static final class Tokenizer {
-    private final CharSequence text;
-    private final Matcher matcher;
-    private String currentToken;
-
-    // The character index within this.text at which the current token begins.
-    private int pos = 0;
-
-    // The line and column numbers of the current token.
-    private int line = 0;
-    private int column = 0;
-
-    // The line and column numbers of the previous token (allows throwing
-    // errors *after* consuming).
-    private int previousLine = 0;
-    private int previousColumn = 0;
-
-    // We use possessive quantifiers (*+ and ++) because otherwise the Java
-    // regex matcher has stack overflows on large inputs.
-    private static final Pattern WHITESPACE =
-      Pattern.compile("(\\s|(#.*$))++", Pattern.MULTILINE);
-    private static final Pattern TOKEN = Pattern.compile(
-      "[a-zA-Z_][0-9a-zA-Z_+-]*+|" +                // an identifier
-      "[.]?[0-9+-][0-9a-zA-Z_.+-]*+|" +             // a number
-      "\"([^\"\n\\\\]|\\\\.)*+(\"|\\\\?$)|" +       // a double-quoted string
-      "\'([^\'\n\\\\]|\\\\.)*+(\'|\\\\?$)",         // a single-quoted string
-      Pattern.MULTILINE);
-
-    private static final Pattern DOUBLE_INFINITY = Pattern.compile(
-      "-?inf(inity)?",
-      Pattern.CASE_INSENSITIVE);
-    private static final Pattern FLOAT_INFINITY = Pattern.compile(
-      "-?inf(inity)?f?",
-      Pattern.CASE_INSENSITIVE);
-    private static final Pattern FLOAT_NAN = Pattern.compile(
-      "nanf?",
-      Pattern.CASE_INSENSITIVE);
-
-    /** Construct a tokenizer that parses tokens from the given text. */
-    private Tokenizer(final CharSequence text) {
-      this.text = text;
-      this.matcher = WHITESPACE.matcher(text);
-      skipWhitespace();
-      nextToken();
-    }
-
-    int getPreviousLine() {
-      return previousLine;
-    }
-
-    int getPreviousColumn() {
-      return previousColumn;
-    }
-
-    int getLine() {
-      return line;
-    }
-
-    int getColumn() {
-      return column;
-    }
-
-    /** Are we at the end of the input? */
-    public boolean atEnd() {
-      return currentToken.length() == 0;
-    }
-
-    /** Advance to the next token. */
-    public void nextToken() {
-      previousLine = line;
-      previousColumn = column;
-
-      // Advance the line counter to the current position.
-      while (pos < matcher.regionStart()) {
-        if (text.charAt(pos) == '\n') {
-          ++line;
-          column = 0;
-        } else {
-          ++column;
-        }
-        ++pos;
-      }
-
-      // Match the next token.
-      if (matcher.regionStart() == matcher.regionEnd()) {
-        // EOF
-        currentToken = "";
-      } else {
-        matcher.usePattern(TOKEN);
-        if (matcher.lookingAt()) {
-          currentToken = matcher.group();
-          matcher.region(matcher.end(), matcher.regionEnd());
-        } else {
-          // Take one character.
-          currentToken = String.valueOf(text.charAt(pos));
-          matcher.region(pos + 1, matcher.regionEnd());
-        }
-
-        skipWhitespace();
-      }
-    }
-
-    /**
-     * Skip over any whitespace so that the matcher region starts at the next
-     * token.
-     */
-    private void skipWhitespace() {
-      matcher.usePattern(WHITESPACE);
-      if (matcher.lookingAt()) {
-        matcher.region(matcher.end(), matcher.regionEnd());
-      }
-    }
-
-    /**
-     * If the next token exactly matches {@code token}, consume it and return
-     * {@code true}.  Otherwise, return {@code false} without doing anything.
-     */
-    public boolean tryConsume(final String token) {
-      if (currentToken.equals(token)) {
-        nextToken();
-        return true;
-      } else {
-        return false;
-      }
-    }
-
-    /**
-     * If the next token exactly matches {@code token}, consume it.  Otherwise,
-     * throw a {@link ParseException}.
-     */
-    public void consume(final String token) throws ParseException {
-      if (!tryConsume(token)) {
-        throw parseException("Expected \"" + token + "\".");
-      }
-    }
-
-    /**
-     * Returns {@code true} if the next token is an integer, but does
-     * not consume it.
-     */
-    public boolean lookingAtInteger() {
-      if (currentToken.length() == 0) {
-        return false;
-      }
-
-      final char c = currentToken.charAt(0);
-      return ('0' <= c && c <= '9')
-          || c == '-' || c == '+';
-    }
-
-    /**
-     * Returns {@code true} if the current token's text is equal to that
-     * specified.
-     */
-    public boolean lookingAt(String text) {
-      return currentToken.equals(text);
-    }
-
-    /**
-     * If the next token is an identifier, consume it and return its value.
-     * Otherwise, throw a {@link ParseException}.
-     */
-    public String consumeIdentifier() throws ParseException {
-      for (int i = 0; i < currentToken.length(); i++) {
-        final char c = currentToken.charAt(i);
-        if (('a' <= c && c <= 'z')
-            || ('A' <= c && c <= 'Z')
-            || ('0' <= c && c <= '9')
-            || (c == '_') || (c == '.')) {
-          // OK
-        } else {
-          throw parseException(
-              "Expected identifier. Found '" + currentToken + "'");
-        }
-      }
-
-      final String result = currentToken;
-      nextToken();
-      return result;
-    }
-
-    /**
-     * If the next token is an identifier, consume it and return {@code true}.
-     * Otherwise, return {@code false} without doing anything.
-     */
-    public boolean tryConsumeIdentifier() {
-      try {
-        consumeIdentifier();
-        return true;
-      } catch (ParseException e) {
-        return false;
-      }
-    }
-
-    /**
-     * If the next token is a 32-bit signed integer, consume it and return its
-     * value.  Otherwise, throw a {@link ParseException}.
-     */
-    public int consumeInt32() throws ParseException {
-      try {
-        final int result = parseInt32(currentToken);
-        nextToken();
-        return result;
-      } catch (NumberFormatException e) {
-        throw integerParseException(e);
-      }
-    }
-
-    /**
-     * If the next token is a 32-bit unsigned integer, consume it and return its
-     * value.  Otherwise, throw a {@link ParseException}.
-     */
-    public int consumeUInt32() throws ParseException {
-      try {
-        final int result = parseUInt32(currentToken);
-        nextToken();
-        return result;
-      } catch (NumberFormatException e) {
-        throw integerParseException(e);
-      }
-    }
-
-    /**
-     * If the next token is a 64-bit signed integer, consume it and return its
-     * value.  Otherwise, throw a {@link ParseException}.
-     */
-    public long consumeInt64() throws ParseException {
-      try {
-        final long result = parseInt64(currentToken);
-        nextToken();
-        return result;
-      } catch (NumberFormatException e) {
-        throw integerParseException(e);
-      }
-    }
-
-    /**
-     * If the next token is a 64-bit signed integer, consume it and return
-     * {@code true}.  Otherwise, return {@code false} without doing anything.
-     */
-    public boolean tryConsumeInt64() {
-      try {
-        consumeInt64();
-        return true;
-      } catch (ParseException e) {
-        return false;
-      }
-    }
-
-    /**
-     * If the next token is a 64-bit unsigned integer, consume it and return its
-     * value.  Otherwise, throw a {@link ParseException}.
-     */
-    public long consumeUInt64() throws ParseException {
-      try {
-        final long result = parseUInt64(currentToken);
-        nextToken();
-        return result;
-      } catch (NumberFormatException e) {
-        throw integerParseException(e);
-      }
-    }
-
-    /**
-     * If the next token is a 64-bit unsigned integer, consume it and return
-     * {@code true}.  Otherwise, return {@code false} without doing anything.
-     */
-    public boolean tryConsumeUInt64() {
-      try {
-        consumeUInt64();
-        return true;
-      } catch (ParseException e) {
-        return false;
-      }
-    }
-
-    /**
-     * If the next token is a double, consume it and return its value.
-     * Otherwise, throw a {@link ParseException}.
-     */
-    public double consumeDouble() throws ParseException {
-      // We need to parse infinity and nan separately because
-      // Double.parseDouble() does not accept "inf", "infinity", or "nan".
-      if (DOUBLE_INFINITY.matcher(currentToken).matches()) {
-        final boolean negative = currentToken.startsWith("-");
-        nextToken();
-        return negative ? Double.NEGATIVE_INFINITY : Double.POSITIVE_INFINITY;
-      }
-      if (currentToken.equalsIgnoreCase("nan")) {
-        nextToken();
-        return Double.NaN;
-      }
-      try {
-        final double result = Double.parseDouble(currentToken);
-        nextToken();
-        return result;
-      } catch (NumberFormatException e) {
-        throw floatParseException(e);
-      }
-    }
-
-    /**
-     * If the next token is a double, consume it and return {@code true}.
-     * Otherwise, return {@code false} without doing anything.
-     */
-    public boolean tryConsumeDouble() {
-      try {
-        consumeDouble();
-        return true;
-      } catch (ParseException e) {
-        return false;
-      }
-    }
-
-    /**
-     * If the next token is a float, consume it and return its value.
-     * Otherwise, throw a {@link ParseException}.
-     */
-    public float consumeFloat() throws ParseException {
-      // We need to parse infinity and nan separately because
-      // Float.parseFloat() does not accept "inf", "infinity", or "nan".
-      if (FLOAT_INFINITY.matcher(currentToken).matches()) {
-        final boolean negative = currentToken.startsWith("-");
-        nextToken();
-        return negative ? Float.NEGATIVE_INFINITY : Float.POSITIVE_INFINITY;
-      }
-      if (FLOAT_NAN.matcher(currentToken).matches()) {
-        nextToken();
-        return Float.NaN;
-      }
-      try {
-        final float result = Float.parseFloat(currentToken);
-        nextToken();
-        return result;
-      } catch (NumberFormatException e) {
-        throw floatParseException(e);
-      }
-    }
-
-    /**
-     * If the next token is a float, consume it and return {@code true}.
-     * Otherwise, return {@code false} without doing anything.
-     */
-    public boolean tryConsumeFloat() {
-      try {
-        consumeFloat();
-        return true;
-      } catch (ParseException e) {
-        return false;
-      }
-    }
-
-    /**
-     * If the next token is a boolean, consume it and return its value.
-     * Otherwise, throw a {@link ParseException}.
-     */
-    public boolean consumeBoolean() throws ParseException {
-      if (currentToken.equals("true")
-          || currentToken.equals("True")
-          || currentToken.equals("t")
-          || currentToken.equals("1")) {
-        nextToken();
-        return true;
-      } else if (currentToken.equals("false")
-          || currentToken.equals("False")
-          || currentToken.equals("f")
-          || currentToken.equals("0")) {
-        nextToken();
-        return false;
-      } else {
-        throw parseException("Expected \"true\" or \"false\".");
-      }
-    }
-
-    /**
-     * If the next token is a string, consume it and return its (unescaped)
-     * value.  Otherwise, throw a {@link ParseException}.
-     */
-    public String consumeString() throws ParseException {
-      return consumeByteString().toStringUtf8();
-    }
-
-    /**
-     * If the next token is a string, consume it and return true.  Otherwise,
-     * return false.
-     */
-    public boolean tryConsumeString() {
-      try {
-        consumeString();
-        return true;
-      } catch (ParseException e) {
-        return false;
-      }
-    }
-
-    /**
-     * If the next token is a string, consume it, unescape it as a
-     * {@link ByteString}, and return it.  Otherwise, throw a
-     * {@link ParseException}.
-     */
-    public ByteString consumeByteString() throws ParseException {
-      List<ByteString> list = new ArrayList<ByteString>();
-      consumeByteString(list);
-      while (currentToken.startsWith("'") || currentToken.startsWith("\"")) {
-        consumeByteString(list);
-      }
-      return ByteString.copyFrom(list);
-    }
-
-    /**
-     * Like {@link #consumeByteString()} but adds each token of the string to
-     * the given list.  String literals (whether bytes or text) may come in
-     * multiple adjacent tokens which are automatically concatenated, like in
-     * C or Python.
-     */
-    private void consumeByteString(List<ByteString> list)
-        throws ParseException {
-      final char quote = currentToken.length() > 0
-          ? currentToken.charAt(0)
-          : '\0';
-      if (quote != '\"' && quote != '\'') {
-        throw parseException("Expected string.");
-      }
-
-      if (currentToken.length() < 2
-          || currentToken.charAt(currentToken.length() - 1) != quote) {
-        throw parseException("String missing ending quote.");
-      }
-
-      try {
-        final String escaped =
-            currentToken.substring(1, currentToken.length() - 1);
-        final ByteString result = unescapeBytes(escaped);
-        nextToken();
-        list.add(result);
-      } catch (InvalidEscapeSequenceException e) {
-        throw parseException(e.getMessage());
-      }
-    }
-
-    /**
-     * Returns a {@link ParseException} with the current line and column
-     * numbers in the description, suitable for throwing.
-     */
-    public ParseException parseException(final String description) {
-      // Note:  People generally prefer one-based line and column numbers.
-      return new ParseException(
-        line + 1, column + 1, description);
-    }
-
-    /**
-     * Returns a {@link ParseException} with the line and column numbers of
-     * the previous token in the description, suitable for throwing.
-     */
-    public ParseException parseExceptionPreviousToken(
-        final String description) {
-      // Note:  People generally prefer one-based line and column numbers.
-      return new ParseException(
-        previousLine + 1, previousColumn + 1, description);
-    }
-
-    /**
-     * Constructs an appropriate {@link ParseException} for the given
-     * {@code NumberFormatException} when trying to parse an integer.
-     */
-    private ParseException integerParseException(
-        final NumberFormatException e) {
-      return parseException("Couldn't parse integer: " + e.getMessage());
-    }
-
-    /**
-     * Constructs an appropriate {@link ParseException} for the given
-     * {@code NumberFormatException} when trying to parse a float or double.
-     */
-    private ParseException floatParseException(final NumberFormatException e) {
-      return parseException("Couldn't parse number: " + e.getMessage());
-    }
-
-    /**
-     * Returns a {@link UnknownFieldParseException} with the line and column
-     * numbers of the previous token in the description, and the unknown field
-     * name, suitable for throwing.
-     */
-    public UnknownFieldParseException unknownFieldParseExceptionPreviousToken(
-        final String unknownField, final String description) {
-      // Note:  People generally prefer one-based line and column numbers.
-      return new UnknownFieldParseException(
-        previousLine + 1, previousColumn + 1, unknownField, description);
-    }
-  }
-
-  /** Thrown when parsing an invalid text format message. */
-  public static class ParseException extends IOException {
-    private static final long serialVersionUID = 3196188060225107702L;
-
-    private final int line;
-    private final int column;
-
-    /** Create a new instance, with -1 as the line and column numbers. */
-    public ParseException(final String message) {
-      this(-1, -1, message);
-    }
-
-    /**
-     * Create a new instance
-     *
-     * @param line the line number where the parse error occurred,
-     * using 1-offset.
-     * @param column the column number where the parser error occurred,
-     * using 1-offset.
-     */
-    public ParseException(final int line, final int column,
-        final String message) {
-      super(Integer.toString(line) + ":" + column + ": " + message);
-      this.line = line;
-      this.column = column;
-    }
-
-    /**
-     * Return the line where the parse exception occurred, or -1 when
-     * none is provided. The value is specified as 1-offset, so the first
-     * line is line 1.
-     */
-    public int getLine() {
-      return line;
-    }
-
-    /**
-     * Return the column where the parse exception occurred, or -1 when
-     * none is provided. The value is specified as 1-offset, so the first
-     * line is line 1.
-     */
-    public int getColumn() {
-      return column;
-    }
-  }
-
-  /**
-   * Thrown when encountering an unknown field while parsing
-   * a text format message.
-   */
-  public static class UnknownFieldParseException extends ParseException {
-    private final String unknownField;
-
-    /**
-     * Create a new instance, with -1 as the line and column numbers, and an
-     * empty unknown field name.
-     */
-    public UnknownFieldParseException(final String message) {
-      this(-1, -1, "", message);
-    }
-
-    /**
-     * Create a new instance
-     *
-     * @param line the line number where the parse error occurred,
-     * using 1-offset.
-     * @param column the column number where the parser error occurred,
-     * using 1-offset.
-     * @param unknownField the name of the unknown field found while parsing.
-     */
-    public UnknownFieldParseException(final int line, final int column,
-        final String unknownField, final String message) {
-      super(line, column, message);
-      this.unknownField = unknownField;
-    }
-
-    /**
-     * Return the name of the unknown field encountered while parsing the
-     * protocol buffer string.
-     */
-    public String getUnknownField() {
-      return unknownField;
-    }
-  }
-
-  private static final Parser PARSER = Parser.newBuilder().build();
-
-  /**
-   * Return a {@link Parser} instance which can parse text-format
-   * messages. The returned instance is thread-safe.
-   */
-  public static Parser getParser() {
-    return PARSER;
-  }
-
-  /**
-   * Parse a text-format message from {@code input} and merge the contents
-   * into {@code builder}.
-   */
-  public static void merge(final Readable input,
-                           final Message.Builder builder)
-                           throws IOException {
-    PARSER.merge(input, builder);
-  }
-
-  /**
-   * Parse a text-format message from {@code input} and merge the contents
-   * into {@code builder}.
-   */
-  public static void merge(final CharSequence input,
-                           final Message.Builder builder)
-                           throws ParseException {
-    PARSER.merge(input, builder);
-  }
-
-  /**
-   * Parse a text-format message from {@code input} and merge the contents
-   * into {@code builder}.  Extensions will be recognized if they are
-   * registered in {@code extensionRegistry}.
-   */
-  public static void merge(final Readable input,
-                           final ExtensionRegistry extensionRegistry,
-                           final Message.Builder builder)
-                           throws IOException {
-    PARSER.merge(input, extensionRegistry, builder);
-  }
-
-
-  /**
-   * Parse a text-format message from {@code input} and merge the contents
-   * into {@code builder}.  Extensions will be recognized if they are
-   * registered in {@code extensionRegistry}.
-   */
-  public static void merge(final CharSequence input,
-                           final ExtensionRegistry extensionRegistry,
-                           final Message.Builder builder)
-                           throws ParseException {
-    PARSER.merge(input, extensionRegistry, builder);
-  }
-
-
-  /**
-   * Parser for text-format proto2 instances. This class is thread-safe.
-   * The implementation largely follows google/protobuf/text_format.cc.
-   *
-   * <p>Use {@link TextFormat#getParser()} to obtain the default parser, or
-   * {@link Builder} to control the parser behavior.
-   */
-  public static class Parser {
-    /**
-     * Determines if repeated values for non-repeated fields and
-     * oneofs are permitted. For example, given required/optional field "foo"
-     * and a oneof containing "baz" and "qux":
-     * <ul>
-     * <li>"foo: 1 foo: 2"
-     * <li>"baz: 1 qux: 2"
-     * <li>merging "foo: 2" into a proto in which foo is already set, or
-     * <li>merging "qux: 2" into a proto in which baz is already set.
-     * </ul>
-     */
-    public enum SingularOverwritePolicy {
-      /** The last value is retained. */
-      ALLOW_SINGULAR_OVERWRITES,
-      /** An error is issued. */
-      FORBID_SINGULAR_OVERWRITES
-    }
-
-    private final boolean allowUnknownFields;
-    private final SingularOverwritePolicy singularOverwritePolicy;
-    private TextFormatParseInfoTree.Builder parseInfoTreeBuilder;
-
-    private Parser(
-        boolean allowUnknownFields, SingularOverwritePolicy singularOverwritePolicy,
-        TextFormatParseInfoTree.Builder parseInfoTreeBuilder) {
-      this.allowUnknownFields = allowUnknownFields;
-      this.singularOverwritePolicy = singularOverwritePolicy;
-      this.parseInfoTreeBuilder = parseInfoTreeBuilder;
-    }
-
-    /**
-     * Returns a new instance of {@link Builder}.
-     */
-    public static Builder newBuilder() {
-      return new Builder();
-    }
-
-    /**
-     * Builder that can be used to obtain new instances of {@link Parser}.
-     */
-    public static class Builder {
-      private boolean allowUnknownFields = false;
-      private SingularOverwritePolicy singularOverwritePolicy =
-          SingularOverwritePolicy.ALLOW_SINGULAR_OVERWRITES;
-      private TextFormatParseInfoTree.Builder parseInfoTreeBuilder = null;
-
-
-      /**
-       * Sets parser behavior when a non-repeated field appears more than once.
-       */
-      public Builder setSingularOverwritePolicy(SingularOverwritePolicy p) {
-        this.singularOverwritePolicy = p;
-        return this;
-      }
-
-      public Builder setParseInfoTreeBuilder(
-          TextFormatParseInfoTree.Builder parseInfoTreeBuilder) {
-        this.parseInfoTreeBuilder = parseInfoTreeBuilder;
-        return this;
-      }
-
-      public Parser build() {
-        return new Parser(
-            allowUnknownFields, singularOverwritePolicy, parseInfoTreeBuilder);
-      }
-    }
-
-    /**
-     * Parse a text-format message from {@code input} and merge the contents
-     * into {@code builder}.
-     */
-    public void merge(final Readable input,
-                      final Message.Builder builder)
-                      throws IOException {
-      merge(input, ExtensionRegistry.getEmptyRegistry(), builder);
-    }
-
-    /**
-     * Parse a text-format message from {@code input} and merge the contents
-     * into {@code builder}.
-     */
-    public void merge(final CharSequence input,
-                      final Message.Builder builder)
-                      throws ParseException {
-      merge(input, ExtensionRegistry.getEmptyRegistry(), builder);
-    }
-
-    /**
-     * Parse a text-format message from {@code input} and merge the contents
-     * into {@code builder}.  Extensions will be recognized if they are
-     * registered in {@code extensionRegistry}.
-     */
-    public void merge(final Readable input,
-                      final ExtensionRegistry extensionRegistry,
-                      final Message.Builder builder)
-                      throws IOException {
-      // Read the entire input to a String then parse that.
-
-      // If StreamTokenizer were not quite so crippled, or if there were a kind
-      // of Reader that could read in chunks that match some particular regex,
-      // or if we wanted to write a custom Reader to tokenize our stream, then
-      // we would not have to read to one big String.  Alas, none of these is
-      // the case.  Oh well.
-
-      merge(toStringBuilder(input), extensionRegistry, builder);
-    }
-
-
-    private static final int BUFFER_SIZE = 4096;
-
-    // TODO(chrisn): See if working around java.io.Reader#read(CharBuffer)
-    // overhead is worthwhile
-    private static StringBuilder toStringBuilder(final Readable input)
-        throws IOException {
-      final StringBuilder text = new StringBuilder();
-      final CharBuffer buffer = CharBuffer.allocate(BUFFER_SIZE);
-      while (true) {
-        final int n = input.read(buffer);
-        if (n == -1) {
-          break;
-        }
-        buffer.flip();
-        text.append(buffer, 0, n);
-      }
-      return text;
-    }
-
-    // Check both unknown fields and unknown extensions and log warming messages
-    // or throw exceptions according to the flag.
-    private void checkUnknownFields(final List<String> unknownFields)
-        throws ParseException {
-      if (unknownFields.isEmpty()) {
-        return;
-      }
-
-      StringBuilder msg = new StringBuilder("Input contains unknown fields and/or extensions:");
-      for (String field : unknownFields) {
-        msg.append('\n').append(field);
-      }
-
-      if (allowUnknownFields) {
-          logger.warning(msg.toString());
-      } else {
-        String[] lineColumn = unknownFields.get(0).split(":");
-        throw new ParseException(Integer.valueOf(lineColumn[0]),
-            Integer.valueOf(lineColumn[1]), msg.toString());
-      }
-    }
-
-    /**
-     * Parse a text-format message from {@code input} and merge the contents
-     * into {@code builder}.  Extensions will be recognized if they are
-     * registered in {@code extensionRegistry}.
-     */
-    public void merge(final CharSequence input,
-                      final ExtensionRegistry extensionRegistry,
-                      final Message.Builder builder)
-                      throws ParseException {
-      final Tokenizer tokenizer = new Tokenizer(input);
-      MessageReflection.BuilderAdapter target =
-          new MessageReflection.BuilderAdapter(builder);
-
-      List<String> unknownFields = new ArrayList<String>();
-
-      while (!tokenizer.atEnd()) {
-        mergeField(tokenizer, extensionRegistry, target, unknownFields);
-      }
-
-      checkUnknownFields(unknownFields);
-    }
-
-
-    /**
-     * Parse a single field from {@code tokenizer} and merge it into
-     * {@code builder}.
-     */
-    private void mergeField(final Tokenizer tokenizer,
-                            final ExtensionRegistry extensionRegistry,
-                            final MessageReflection.MergeTarget target,
-                            List<String> unknownFields)
-                            throws ParseException {
-      mergeField(tokenizer, extensionRegistry, target, parseInfoTreeBuilder,
-                 unknownFields);
-    }
-
-    /**
-     * Parse a single field from {@code tokenizer} and merge it into
-     * {@code builder}.
-     */
-    private void mergeField(final Tokenizer tokenizer,
-                            final ExtensionRegistry extensionRegistry,
-                            final MessageReflection.MergeTarget target,
-                            TextFormatParseInfoTree.Builder parseTreeBuilder,
-                            List<String> unknownFields)
-                            throws ParseException {
-      FieldDescriptor field = null;
-      int startLine = tokenizer.getLine();
-      int startColumn = tokenizer.getColumn();
-      final Descriptor type = target.getDescriptorForType();
-      ExtensionRegistry.ExtensionInfo extension = null;
-
-      if (tokenizer.tryConsume("[")) {
-        // An extension.
-        final StringBuilder name =
-            new StringBuilder(tokenizer.consumeIdentifier());
-        while (tokenizer.tryConsume(".")) {
-          name.append('.');
-          name.append(tokenizer.consumeIdentifier());
-        }
-
-        extension = target.findExtensionByName(
-            extensionRegistry, name.toString());
-
-        if (extension == null) {
-          unknownFields.add((tokenizer.getPreviousLine() + 1) + ":" +
-              (tokenizer.getPreviousColumn() + 1) + ":\t" +
-              type.getFullName() + ".[" + name + "]");
-        } else {
-          if (extension.descriptor.getContainingType() != type) {
-            throw tokenizer.parseExceptionPreviousToken(
-              "Extension \"" + name + "\" does not extend message type \""
-              + type.getFullName() + "\".");
-          }
-          field = extension.descriptor;
-        }
-
-        tokenizer.consume("]");
-      } else {
-        final String name = tokenizer.consumeIdentifier();
-        field = type.findFieldByName(name);
-
-        // Group names are expected to be capitalized as they appear in the
-        // .proto file, which actually matches their type names, not their field
-        // names.
-        if (field == null) {
-          // Explicitly specify US locale so that this code does not break when
-          // executing in Turkey.
-          final String lowerName = name.toLowerCase(Locale.US);
-          field = type.findFieldByName(lowerName);
-          // If the case-insensitive match worked but the field is NOT a group,
-          if (field != null && field.getType() != FieldDescriptor.Type.GROUP) {
-            field = null;
-          }
-        }
-        // Again, special-case group names as described above.
-        if (field != null && field.getType() == FieldDescriptor.Type.GROUP
-            && !field.getMessageType().getName().equals(name)) {
-          field = null;
-        }
-
-        if (field == null) {
-          unknownFields.add((tokenizer.getPreviousLine() + 1) + ":" +
-              (tokenizer.getPreviousColumn() + 1) + ":\t" +
-              type.getFullName() + "." + name);
-        }
-      }
-
-      // Skips unknown fields.
-      if (field == null) {
-        // Try to guess the type of this field.
-        // If this field is not a message, there should be a ":" between the
-        // field name and the field value and also the field value should not
-        // start with "{" or "<" which indicates the beginning of a message body.
-        // If there is no ":" or there is a "{" or "<" after ":", this field has
-        // to be a message or the input is ill-formed.
-        if (tokenizer.tryConsume(":")
-            && !tokenizer.lookingAt("{")
-            && !tokenizer.lookingAt("<")) {
-          skipFieldValue(tokenizer);
-        } else {
-          skipFieldMessage(tokenizer);
-        }
-        return;
-      }
-
-      // Handle potential ':'.
-      if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-        tokenizer.tryConsume(":");  // optional
-        if (parseTreeBuilder != null) {
-          TextFormatParseInfoTree.Builder childParseTreeBuilder =
-              parseTreeBuilder.getBuilderForSubMessageField(field);
-          consumeFieldValues(tokenizer, extensionRegistry, target, field, extension,
-              childParseTreeBuilder, unknownFields);
-        } else {
-          consumeFieldValues(tokenizer, extensionRegistry, target, field, extension,
-              parseTreeBuilder, unknownFields);
-        }
-      } else {
-        tokenizer.consume(":");  // required
-        consumeFieldValues(tokenizer, extensionRegistry, target, field,
-            extension, parseTreeBuilder, unknownFields);
-      }
-
-      if (parseTreeBuilder != null) {
-        parseTreeBuilder.setLocation(
-            field, TextFormatParseLocation.create(startLine, startColumn));
-      }
-
-      // For historical reasons, fields may optionally be separated by commas or
-      // semicolons.
-      if (!tokenizer.tryConsume(";")) {
-        tokenizer.tryConsume(",");
-      }
-    }
-
-    /**
-     * Parse a one or more field values from {@code tokenizer} and merge it into
-     * {@code builder}.
-     */
-    private void consumeFieldValues(
-        final Tokenizer tokenizer,
-        final ExtensionRegistry extensionRegistry,
-        final MessageReflection.MergeTarget target,
-        final FieldDescriptor field,
-        final ExtensionRegistry.ExtensionInfo extension,
-        final TextFormatParseInfoTree.Builder parseTreeBuilder,
-        List<String> unknownFields)
-        throws ParseException {
-      // Support specifying repeated field values as a comma-separated list.
-      // Ex."foo: [1, 2, 3]"
-      if (field.isRepeated() && tokenizer.tryConsume("[")) {
-        if (!tokenizer.tryConsume("]")) {  // Allow "foo: []" to be treated as empty.
-          while (true) {
-            consumeFieldValue(
-                tokenizer,
-                extensionRegistry,
-                target,
-                field,
-                extension,
-                parseTreeBuilder,
-                unknownFields);
-            if (tokenizer.tryConsume("]")) {
-              // End of list.
-              break;
-            }
-            tokenizer.consume(",");
-          }
-        }
-      } else {
-        consumeFieldValue(tokenizer, extensionRegistry, target, field,
-            extension, parseTreeBuilder, unknownFields);
-      }
-    }
-
-    /**
-     * Parse a single field value from {@code tokenizer} and merge it into
-     * {@code builder}.
-     */
-    private void consumeFieldValue(
-        final Tokenizer tokenizer,
-        final ExtensionRegistry extensionRegistry,
-        final MessageReflection.MergeTarget target,
-        final FieldDescriptor field,
-        final ExtensionRegistry.ExtensionInfo extension,
-        final TextFormatParseInfoTree.Builder parseTreeBuilder,
-        List<String> unknownFields)
-        throws ParseException {
-      Object value = null;
-
-      if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-        final String endToken;
-        if (tokenizer.tryConsume("<")) {
-          endToken = ">";
-        } else {
-          tokenizer.consume("{");
-          endToken = "}";
-        }
-
-        final MessageReflection.MergeTarget subField;
-        subField = target.newMergeTargetForField(field,
-            (extension == null) ? null : extension.defaultInstance);
-
-        while (!tokenizer.tryConsume(endToken)) {
-          if (tokenizer.atEnd()) {
-            throw tokenizer.parseException(
-              "Expected \"" + endToken + "\".");
-          }
-          mergeField(tokenizer, extensionRegistry, subField, parseTreeBuilder,
-              unknownFields);
-        }
-
-        value = subField.finish();
-
-      } else {
-        switch (field.getType()) {
-          case INT32:
-          case SINT32:
-          case SFIXED32:
-            value = tokenizer.consumeInt32();
-            break;
-
-          case INT64:
-          case SINT64:
-          case SFIXED64:
-            value = tokenizer.consumeInt64();
-            break;
-
-          case UINT32:
-          case FIXED32:
-            value = tokenizer.consumeUInt32();
-            break;
-
-          case UINT64:
-          case FIXED64:
-            value = tokenizer.consumeUInt64();
-            break;
-
-          case FLOAT:
-            value = tokenizer.consumeFloat();
-            break;
-
-          case DOUBLE:
-            value = tokenizer.consumeDouble();
-            break;
-
-          case BOOL:
-            value = tokenizer.consumeBoolean();
-            break;
-
-          case STRING:
-            value = tokenizer.consumeString();
-            break;
-
-          case BYTES:
-            value = tokenizer.consumeByteString();
-            break;
-
-          case ENUM:
-            final EnumDescriptor enumType = field.getEnumType();
-
-            if (tokenizer.lookingAtInteger()) {
-              final int number = tokenizer.consumeInt32();
-              value = enumType.findValueByNumber(number);
-              if (value == null) {
-                throw tokenizer.parseExceptionPreviousToken(
-                  "Enum type \"" + enumType.getFullName()
-                  + "\" has no value with number " + number + '.');
-              }
-            } else {
-              final String id = tokenizer.consumeIdentifier();
-              value = enumType.findValueByName(id);
-              if (value == null) {
-                throw tokenizer.parseExceptionPreviousToken(
-                  "Enum type \"" + enumType.getFullName()
-                  + "\" has no value named \"" + id + "\".");
-              }
-            }
-
-            break;
-
-          case MESSAGE:
-          case GROUP:
-            throw new RuntimeException("Can't get here.");
-        }
-      }
-
-      if (field.isRepeated()) {
-        target.addRepeatedField(field, value);
-      } else if ((singularOverwritePolicy
-              == SingularOverwritePolicy.FORBID_SINGULAR_OVERWRITES)
-          && target.hasField(field)) {
-        throw tokenizer.parseExceptionPreviousToken("Non-repeated field \""
-            + field.getFullName() + "\" cannot be overwritten.");
-      } else if ((singularOverwritePolicy
-              == SingularOverwritePolicy.FORBID_SINGULAR_OVERWRITES)
-          && field.getContainingOneof() != null
-          && target.hasOneof(field.getContainingOneof())) {
-        Descriptors.OneofDescriptor oneof = field.getContainingOneof();
-        throw tokenizer.parseExceptionPreviousToken("Field \""
-            + field.getFullName() + "\" is specified along with field \""
-            + target.getOneofFieldDescriptor(oneof).getFullName()
-            + "\", another member of oneof \"" + oneof.getName() + "\".");
-      } else {
-        target.setField(field, value);
-      }
-    }
-
-    /**
-     * Skips the next field including the field's name and value.
-     */
-    private void skipField(Tokenizer tokenizer) throws ParseException {
-      if (tokenizer.tryConsume("[")) {
-        // Extension name.
-        do {
-          tokenizer.consumeIdentifier();
-        } while (tokenizer.tryConsume("."));
-        tokenizer.consume("]");
-      } else {
-        tokenizer.consumeIdentifier();
-      }
-
-      // Try to guess the type of this field.
-      // If this field is not a message, there should be a ":" between the
-      // field name and the field value and also the field value should not
-      // start with "{" or "<" which indicates the beginning of a message body.
-      // If there is no ":" or there is a "{" or "<" after ":", this field has
-      // to be a message or the input is ill-formed.
-      if (tokenizer.tryConsume(":")
-          && !tokenizer.lookingAt("<")
-          && !tokenizer.lookingAt("{")) {
-        skipFieldValue(tokenizer);
-      } else {
-        skipFieldMessage(tokenizer);
-      }
-      // For historical reasons, fields may optionally be separated by commas or
-      // semicolons.
-      if (!tokenizer.tryConsume(";")) {
-        tokenizer.tryConsume(",");
-      }
-    }
-
-    /**
-     * Skips the whole body of a message including the beginning delimiter and
-     * the ending delimiter.
-     */
-    private void skipFieldMessage(Tokenizer tokenizer) throws ParseException {
-      final String delimiter;
-      if (tokenizer.tryConsume("<")) {
-        delimiter = ">";
-      } else {
-        tokenizer.consume("{");
-        delimiter = "}";
-      }
-      while (!tokenizer.lookingAt(">") && !tokenizer.lookingAt("}")) {
-        skipField(tokenizer);
-      }
-      tokenizer.consume(delimiter);
-    }
-
-    /**
-     * Skips a field value.
-     */
-    private void skipFieldValue(Tokenizer tokenizer) throws ParseException {
-      if (tokenizer.tryConsumeString()) {
-        while (tokenizer.tryConsumeString()) {}
-        return;
-      }
-      if (!tokenizer.tryConsumeIdentifier()   // includes enum & boolean
-          && !tokenizer.tryConsumeInt64()     // includes int32
-          && !tokenizer.tryConsumeUInt64()    // includes uint32
-          && !tokenizer.tryConsumeDouble()
-          && !tokenizer.tryConsumeFloat()) {
-        throw tokenizer.parseException(
-            "Invalid field value: " + tokenizer.currentToken);
-      }
-    }
-  }
-
-  // =================================================================
-  // Utility functions
-  //
-  // Some of these methods are package-private because Descriptors.java uses
-  // them.
-
-  /**
-   * Escapes bytes in the format used in protocol buffer text format, which
-   * is the same as the format used for C string literals.  All bytes
-   * that are not printable 7-bit ASCII characters are escaped, as well as
-   * backslash, single-quote, and double-quote characters.  Characters for
-   * which no defined short-hand escape sequence is defined will be escaped
-   * using 3-digit octal sequences.
-   */
-  public static String escapeBytes(ByteString input) {
-    return TextFormatEscaper.escapeBytes(input);
-  }
-
-  /**
-   * Like {@link #escapeBytes(ByteString)}, but used for byte array.
-   */
-  public static String escapeBytes(byte[] input) {
-    return TextFormatEscaper.escapeBytes(input);
-  }
-
-  /**
-   * Un-escape a byte sequence as escaped using
-   * {@link #escapeBytes(ByteString)}.  Two-digit hex escapes (starting with
-   * "\x") are also recognized.
-   */
-  public static ByteString unescapeBytes(final CharSequence charString)
-      throws InvalidEscapeSequenceException {
-    // First convert the Java character sequence to UTF-8 bytes.
-    ByteString input = ByteString.copyFromUtf8(charString.toString());
-    // Then unescape certain byte sequences introduced by ASCII '\\'.  The valid
-    // escapes can all be expressed with ASCII characters, so it is safe to
-    // operate on bytes here.
-    //
-    // Unescaping the input byte array will result in a byte sequence that's no
-    // longer than the input.  That's because each escape sequence is between
-    // two and four bytes long and stands for a single byte.
-    final byte[] result = new byte[input.size()];
-    int pos = 0;
-    for (int i = 0; i < input.size(); i++) {
-      byte c = input.byteAt(i);
-      if (c == '\\') {
-        if (i + 1 < input.size()) {
-          ++i;
-          c = input.byteAt(i);
-          if (isOctal(c)) {
-            // Octal escape.
-            int code = digitValue(c);
-            if (i + 1 < input.size() && isOctal(input.byteAt(i + 1))) {
-              ++i;
-              code = code * 8 + digitValue(input.byteAt(i));
-            }
-            if (i + 1 < input.size() && isOctal(input.byteAt(i + 1))) {
-              ++i;
-              code = code * 8 + digitValue(input.byteAt(i));
-            }
-            // TODO: Check that 0 <= code && code <= 0xFF.
-            result[pos++] = (byte) code;
-          } else {
-            switch (c) {
-              case 'a' : result[pos++] = 0x07; break;
-              case 'b' : result[pos++] = '\b'; break;
-              case 'f' : result[pos++] = '\f'; break;
-              case 'n' : result[pos++] = '\n'; break;
-              case 'r' : result[pos++] = '\r'; break;
-              case 't' : result[pos++] = '\t'; break;
-              case 'v' : result[pos++] = 0x0b; break;
-              case '\\': result[pos++] = '\\'; break;
-              case '\'': result[pos++] = '\''; break;
-              case '"' : result[pos++] = '\"'; break;
-
-              case 'x':
-                // hex escape
-                int code = 0;
-                if (i + 1 < input.size() && isHex(input.byteAt(i + 1))) {
-                  ++i;
-                  code = digitValue(input.byteAt(i));
-                } else {
-                  throw new InvalidEscapeSequenceException(
-                      "Invalid escape sequence: '\\x' with no digits");
-                }
-                if (i + 1 < input.size() && isHex(input.byteAt(i + 1))) {
-                  ++i;
-                  code = code * 16 + digitValue(input.byteAt(i));
-                }
-                result[pos++] = (byte) code;
-                break;
-
-              default:
-                throw new InvalidEscapeSequenceException(
-                    "Invalid escape sequence: '\\" + (char) c + '\'');
-            }
-          }
-        } else {
-          throw new InvalidEscapeSequenceException(
-              "Invalid escape sequence: '\\' at end of string.");
-        }
-      } else {
-        result[pos++] = c;
-      }
-    }
-
-    return result.length == pos
-        ? ByteString.wrap(result)  // This reference has not been out of our control.
-        : ByteString.copyFrom(result, 0, pos);
-  }
-
-  /**
-   * Thrown by {@link TextFormat#unescapeBytes} and
-   * {@link TextFormat#unescapeText} when an invalid escape sequence is seen.
-   */
-  public static class InvalidEscapeSequenceException extends IOException {
-    private static final long serialVersionUID = -8164033650142593304L;
-
-    InvalidEscapeSequenceException(final String description) {
-      super(description);
-    }
-  }
-
-  /**
-   * Like {@link #escapeBytes(ByteString)}, but escapes a text string.
-   * Non-ASCII characters are first encoded as UTF-8, then each byte is escaped
-   * individually as a 3-digit octal escape.  Yes, it's weird.
-   */
-  static String escapeText(final String input) {
-    return escapeBytes(ByteString.copyFromUtf8(input));
-  }
-
-  /**
-   * Escape double quotes and backslashes in a String for unicode output of a message.
-   */
-  public static String escapeDoubleQuotesAndBackslashes(final String input) {
-    return TextFormatEscaper.escapeDoubleQuotesAndBackslashes(input);
-  }
-
-  /**
-   * Un-escape a text string as escaped using {@link #escapeText(String)}.
-   * Two-digit hex escapes (starting with "\x") are also recognized.
-   */
-  static String unescapeText(final String input)
-                             throws InvalidEscapeSequenceException {
-    return unescapeBytes(input).toStringUtf8();
-  }
-
-  /** Is this an octal digit? */
-  private static boolean isOctal(final byte c) {
-    return '0' <= c && c <= '7';
-  }
-
-  /** Is this a hex digit? */
-  private static boolean isHex(final byte c) {
-    return ('0' <= c && c <= '9')
-        || ('a' <= c && c <= 'f')
-        || ('A' <= c && c <= 'F');
-  }
-
-  /**
-   * Interpret a character as a digit (in any base up to 36) and return the
-   * numeric value.  This is like {@code Character.digit()} but we don't accept
-   * non-ASCII digits.
-   */
-  private static int digitValue(final byte c) {
-    if ('0' <= c && c <= '9') {
-      return c - '0';
-    } else if ('a' <= c && c <= 'z') {
-      return c - 'a' + 10;
-    } else {
-      return c - 'A' + 10;
-    }
-  }
-
-  /**
-   * Parse a 32-bit signed integer from the text.  Unlike the Java standard
-   * {@code Integer.parseInt()}, this function recognizes the prefixes "0x"
-   * and "0" to signify hexadecimal and octal numbers, respectively.
-   */
-  static int parseInt32(final String text) throws NumberFormatException {
-    return (int) parseInteger(text, true, false);
-  }
-
-  /**
-   * Parse a 32-bit unsigned integer from the text.  Unlike the Java standard
-   * {@code Integer.parseInt()}, this function recognizes the prefixes "0x"
-   * and "0" to signify hexadecimal and octal numbers, respectively.  The
-   * result is coerced to a (signed) {@code int} when returned since Java has
-   * no unsigned integer type.
-   */
-  static int parseUInt32(final String text) throws NumberFormatException {
-    return (int) parseInteger(text, false, false);
-  }
-
-  /**
-   * Parse a 64-bit signed integer from the text.  Unlike the Java standard
-   * {@code Integer.parseInt()}, this function recognizes the prefixes "0x"
-   * and "0" to signify hexadecimal and octal numbers, respectively.
-   */
-  static long parseInt64(final String text) throws NumberFormatException {
-    return parseInteger(text, true, true);
-  }
-
-  /**
-   * Parse a 64-bit unsigned integer from the text.  Unlike the Java standard
-   * {@code Integer.parseInt()}, this function recognizes the prefixes "0x"
-   * and "0" to signify hexadecimal and octal numbers, respectively.  The
-   * result is coerced to a (signed) {@code long} when returned since Java has
-   * no unsigned long type.
-   */
-  static long parseUInt64(final String text) throws NumberFormatException {
-    return parseInteger(text, false, true);
-  }
-
-  private static long parseInteger(final String text,
-                                   final boolean isSigned,
-                                   final boolean isLong)
-                                   throws NumberFormatException {
-    int pos = 0;
-
-    boolean negative = false;
-    if (text.startsWith("-", pos)) {
-      if (!isSigned) {
-        throw new NumberFormatException("Number must be positive: " + text);
-      }
-      ++pos;
-      negative = true;
-    }
-
-    int radix = 10;
-    if (text.startsWith("0x", pos)) {
-      pos += 2;
-      radix = 16;
-    } else if (text.startsWith("0", pos)) {
-      radix = 8;
-    }
-
-    final String numberText = text.substring(pos);
-
-    long result = 0;
-    if (numberText.length() < 16) {
-      // Can safely assume no overflow.
-      result = Long.parseLong(numberText, radix);
-      if (negative) {
-        result = -result;
-      }
-
-      // Check bounds.
-      // No need to check for 64-bit numbers since they'd have to be 16 chars
-      // or longer to overflow.
-      if (!isLong) {
-        if (isSigned) {
-          if (result > Integer.MAX_VALUE || result < Integer.MIN_VALUE) {
-            throw new NumberFormatException(
-              "Number out of range for 32-bit signed integer: " + text);
-          }
-        } else {
-          if (result >= (1L << 32) || result < 0) {
-            throw new NumberFormatException(
-              "Number out of range for 32-bit unsigned integer: " + text);
-          }
-        }
-      }
-    } else {
-      BigInteger bigValue = new BigInteger(numberText, radix);
-      if (negative) {
-        bigValue = bigValue.negate();
-      }
-
-      // Check bounds.
-      if (!isLong) {
-        if (isSigned) {
-          if (bigValue.bitLength() > 31) {
-            throw new NumberFormatException(
-              "Number out of range for 32-bit signed integer: " + text);
-          }
-        } else {
-          if (bigValue.bitLength() > 32) {
-            throw new NumberFormatException(
-              "Number out of range for 32-bit unsigned integer: " + text);
-          }
-        }
-      } else {
-        if (isSigned) {
-          if (bigValue.bitLength() > 63) {
-            throw new NumberFormatException(
-              "Number out of range for 64-bit signed integer: " + text);
-          }
-        } else {
-          if (bigValue.bitLength() > 64) {
-            throw new NumberFormatException(
-              "Number out of range for 64-bit unsigned integer: " + text);
-          }
-        }
-      }
-
-      result = bigValue.longValue();
-    }
-
-    return result;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatEscaper.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatEscaper.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatEscaper.java
deleted file mode 100644
index 7c50c5d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatEscaper.java
+++ /dev/null
@@ -1,137 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Provide text format escaping support for proto2 instances.
- */
-final class TextFormatEscaper {
-  private TextFormatEscaper() {}
-
-  private interface ByteSequence {
-    int size();
-    byte byteAt(int offset);
-  }
-
-  /**
-   * Escapes bytes in the format used in protocol buffer text format, which
-   * is the same as the format used for C string literals.  All bytes
-   * that are not printable 7-bit ASCII characters are escaped, as well as
-   * backslash, single-quote, and double-quote characters.  Characters for
-   * which no defined short-hand escape sequence is defined will be escaped
-   * using 3-digit octal sequences.
-   */
-  static String escapeBytes(final ByteSequence input) {
-    final StringBuilder builder = new StringBuilder(input.size());
-    for (int i = 0; i < input.size(); i++) {
-      final byte b = input.byteAt(i);
-      switch (b) {
-        // Java does not recognize \a or \v, apparently.
-        case 0x07: builder.append("\\a"); break;
-        case '\b': builder.append("\\b"); break;
-        case '\f': builder.append("\\f"); break;
-        case '\n': builder.append("\\n"); break;
-        case '\r': builder.append("\\r"); break;
-        case '\t': builder.append("\\t"); break;
-        case 0x0b: builder.append("\\v"); break;
-        case '\\': builder.append("\\\\"); break;
-        case '\'': builder.append("\\\'"); break;
-        case '"' : builder.append("\\\""); break;
-        default:
-          // Only ASCII characters between 0x20 (space) and 0x7e (tilde) are
-          // printable.  Other byte values must be escaped.
-          if (b >= 0x20 && b <= 0x7e) {
-            builder.append((char) b);
-          } else {
-            builder.append('\\');
-            builder.append((char) ('0' + ((b >>> 6) & 3)));
-            builder.append((char) ('0' + ((b >>> 3) & 7)));
-            builder.append((char) ('0' + (b & 7)));
-          }
-          break;
-      }
-    }
-    return builder.toString();
-  }
-
-  /**
-   * Escapes bytes in the format used in protocol buffer text format, which
-   * is the same as the format used for C string literals.  All bytes
-   * that are not printable 7-bit ASCII characters are escaped, as well as
-   * backslash, single-quote, and double-quote characters.  Characters for
-   * which no defined short-hand escape sequence is defined will be escaped
-   * using 3-digit octal sequences.
-   */
-  static String escapeBytes(final ByteString input) {
-    return escapeBytes(new ByteSequence() {
-      @Override
-      public int size() {
-        return input.size();
-      }
-      @Override
-      public byte byteAt(int offset) {
-        return input.byteAt(offset);
-      }
-    });
-  }
-
-  /**
-   * Like {@link #escapeBytes(ByteString)}, but used for byte array.
-   */
-  static String escapeBytes(final byte[] input) {
-    return escapeBytes(new ByteSequence() {
-      @Override
-      public int size() {
-        return input.length;
-      }
-      @Override
-      public byte byteAt(int offset) {
-        return input[offset];
-      }
-    });
-  }
-
-  /**
-   * Like {@link #escapeBytes(ByteString)}, but escapes a text string.
-   * Non-ASCII characters are first encoded as UTF-8, then each byte is escaped
-   * individually as a 3-digit octal escape.  Yes, it's weird.
-   */
-  static String escapeText(final String input) {
-    return escapeBytes(ByteString.copyFromUtf8(input));
-  }
-
-  /**
-   * Escape double quotes and backslashes in a String for unicode output of a message.
-   */
-  static String escapeDoubleQuotesAndBackslashes(final String input) {
-    return input.replace("\\", "\\\\").replace("\"", "\\\"");
-  }
-}


[24/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessageV3.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessageV3.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessageV3.java
deleted file mode 100644
index cc815fc..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessageV3.java
+++ /dev/null
@@ -1,2838 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor;
-// In opensource protobuf, we have versioned this GeneratedMessageV3 class to GeneratedMessageV3V3 and
-// in the future may have GeneratedMessageV3V4 etc. This allows us to change some aspects of this
-// class without breaking binary compatibility with old generated code that still subclasses
-// the old GeneratedMessageV3 class. To allow these different GeneratedMessageV3V? classes to
-// interoperate (e.g., a GeneratedMessageV3V3 object has a message extension field whose class
-// type is GeneratedMessageV3V4), these classes still share a common parent class AbstarctMessage
-// and are using the same GeneratedMessage.GeneratedExtension class for extension definitions.
-// Since this class becomes GeneratedMessageV3V? in opensource, we have to add an import here
-// to be able to use GeneratedMessage.GeneratedExtension. The GeneratedExtension definition in
-// this file is also excluded from opensource to avoid conflict.
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessage.GeneratedExtension;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.ObjectStreamException;
-import java.io.Serializable;
-import java.lang.reflect.InvocationTargetException;
-import java.lang.reflect.Method;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-import java.util.TreeMap;
-
-/**
- * All generated protocol message classes extend this class.  This class
- * implements most of the Message and Builder interfaces using Java reflection.
- * Users can ignore this class and pretend that generated messages implement
- * the Message interface directly.
- *
- * @author kenton@google.com Kenton Varda
- */
-public abstract class GeneratedMessageV3 extends AbstractMessage
-    implements Serializable {
-  private static final long serialVersionUID = 1L;
-
-  /**
-   * For testing. Allows a test to disable the optimization that avoids using
-   * field builders for nested messages until they are requested. By disabling
-   * this optimization, existing tests can be reused to test the field builders.
-   */
-  protected static boolean alwaysUseFieldBuilders = false;
-
-  /** For use by generated code only.  */
-  protected UnknownFieldSet unknownFields;
-
-  protected GeneratedMessageV3() {
-    unknownFields = UnknownFieldSet.getDefaultInstance();
-  }
-
-  protected GeneratedMessageV3(Builder<?> builder) {
-    unknownFields = builder.getUnknownFields();
-  }
-
-  @Override
-  public Parser<? extends GeneratedMessageV3> getParserForType() {
-    throw new UnsupportedOperationException(
-        "This is supposed to be overridden by subclasses.");
-  }
-
- /**
-  * For testing. Allows a test to disable the optimization that avoids using
-  * field builders for nested messages until they are requested. By disabling
-  * this optimization, existing tests can be reused to test the field builders.
-  * See {@link RepeatedFieldBuilder} and {@link SingleFieldBuilder}.
-  */
-  static void enableAlwaysUseFieldBuildersForTesting() {
-    alwaysUseFieldBuilders = true;
-  }
-
-  /**
-   * Get the FieldAccessorTable for this type.  We can't have the message
-   * class pass this in to the constructor because of bootstrapping trouble
-   * with DescriptorProtos.
-   */
-  protected abstract FieldAccessorTable internalGetFieldAccessorTable();
-
-  @Override
-  public Descriptor getDescriptorForType() {
-    return internalGetFieldAccessorTable().descriptor;
-  }
-
-  /**
-   * Internal helper to return a modifiable map containing all the fields.
-   * The returned Map is modifialbe so that the caller can add additional
-   * extension fields to implement {@link #getAllFields()}.
-   *
-   * @param getBytesForString whether to generate ByteString for string fields
-   */
-  private Map<FieldDescriptor, Object> getAllFieldsMutable(
-      boolean getBytesForString) {
-    final TreeMap<FieldDescriptor, Object> result =
-      new TreeMap<FieldDescriptor, Object>();
-    final Descriptor descriptor = internalGetFieldAccessorTable().descriptor;
-    final List<FieldDescriptor> fields = descriptor.getFields();
-
-    for (int i = 0; i < fields.size(); i++) {
-      FieldDescriptor field = fields.get(i);
-      final OneofDescriptor oneofDescriptor = field.getContainingOneof();
-
-      /*
-       * If the field is part of a Oneof, then at maximum one field in the Oneof is set
-       * and it is not repeated. There is no need to iterate through the others.
-       */
-      if (oneofDescriptor != null) {
-        // Skip other fields in the Oneof we know are not set
-        i += oneofDescriptor.getFieldCount() - 1;
-        if (!hasOneof(oneofDescriptor)) {
-          // If no field is set in the Oneof, skip all the fields in the Oneof
-          continue;
-        }
-        // Get the pointer to the only field which is set in the Oneof
-        field = getOneofFieldDescriptor(oneofDescriptor);
-      } else {
-        // If we are not in a Oneof, we need to check if the field is set and if it is repeated
-        if (field.isRepeated()) {
-          final List<?> value = (List<?>) getField(field);
-          if (!value.isEmpty()) {
-            result.put(field, value);
-          }
-          continue;
-        }
-        if (!hasField(field)) {
-          continue;
-        }
-      }
-      // Add the field to the map
-      if (getBytesForString && field.getJavaType() == FieldDescriptor.JavaType.STRING) {
-        result.put(field, getFieldRaw(field));
-      } else {
-        result.put(field, getField(field));
-      }
-    }
-    return result;
-  }
-
-  @Override
-  public boolean isInitialized() {
-    for (final FieldDescriptor field : getDescriptorForType().getFields()) {
-      // Check that all required fields are present.
-      if (field.isRequired()) {
-        if (!hasField(field)) {
-          return false;
-        }
-      }
-      // Check that embedded messages are initialized.
-      if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-        if (field.isRepeated()) {
-          @SuppressWarnings("unchecked") final
-          List<Message> messageList = (List<Message>) getField(field);
-          for (final Message element : messageList) {
-            if (!element.isInitialized()) {
-              return false;
-            }
-          }
-        } else {
-          if (hasField(field) && !((Message) getField(field)).isInitialized()) {
-            return false;
-          }
-        }
-      }
-    }
-
-    return true;
-  }
-
-  @Override
-  public Map<FieldDescriptor, Object> getAllFields() {
-    return Collections.unmodifiableMap(
-        getAllFieldsMutable(/* getBytesForString = */ false));
-  }
-
-  /**
-   * Returns a collection of all the fields in this message which are set
-   * and their corresponding values.  A singular ("required" or "optional")
-   * field is set iff hasField() returns true for that field.  A "repeated"
-   * field is set iff getRepeatedFieldCount() is greater than zero.  The
-   * values are exactly what would be returned by calling
-   * {@link #getFieldRaw(Descriptors.FieldDescriptor)} for each field.  The map
-   * is guaranteed to be a sorted map, so iterating over it will return fields
-   * in order by field number.
-   */
-  Map<FieldDescriptor, Object> getAllFieldsRaw() {
-    return Collections.unmodifiableMap(
-        getAllFieldsMutable(/* getBytesForString = */ true));
-  }
-
-  @Override
-  public boolean hasOneof(final OneofDescriptor oneof) {
-    return internalGetFieldAccessorTable().getOneof(oneof).has(this);
-  }
-
-  @Override
-  public FieldDescriptor getOneofFieldDescriptor(final OneofDescriptor oneof) {
-    return internalGetFieldAccessorTable().getOneof(oneof).get(this);
-  }
-
-  @Override
-  public boolean hasField(final FieldDescriptor field) {
-    return internalGetFieldAccessorTable().getField(field).has(this);
-  }
-
-  @Override
-  public Object getField(final FieldDescriptor field) {
-    return internalGetFieldAccessorTable().getField(field).get(this);
-  }
-
-  /**
-   * Obtains the value of the given field, or the default value if it is
-   * not set.  For primitive fields, the boxed primitive value is returned.
-   * For enum fields, the EnumValueDescriptor for the value is returned. For
-   * embedded message fields, the sub-message is returned.  For repeated
-   * fields, a java.util.List is returned. For present string fields, a
-   * ByteString is returned representing the bytes that the field contains.
-   */
-  Object getFieldRaw(final FieldDescriptor field) {
-    return internalGetFieldAccessorTable().getField(field).getRaw(this);
-  }
-
-  @Override
-  public int getRepeatedFieldCount(final FieldDescriptor field) {
-    return internalGetFieldAccessorTable().getField(field)
-      .getRepeatedCount(this);
-  }
-
-  @Override
-  public Object getRepeatedField(final FieldDescriptor field, final int index) {
-    return internalGetFieldAccessorTable().getField(field)
-      .getRepeated(this, index);
-  }
-
-  @Override
-  public UnknownFieldSet getUnknownFields() {
-    throw new UnsupportedOperationException(
-        "This is supposed to be overridden by subclasses.");
-  }
-
-  /**
-   * Called by subclasses to parse an unknown field.
-   * @return {@code true} unless the tag is an end-group tag.
-   */
-  protected boolean parseUnknownField(
-      CodedInputStream input,
-      UnknownFieldSet.Builder unknownFields,
-      ExtensionRegistryLite extensionRegistry,
-      int tag) throws IOException {
-    return unknownFields.mergeFieldFrom(tag, input);
-  }
-
-  protected static <M extends Message> M parseWithIOException(Parser<M> parser, InputStream input)
-      throws IOException {
-    try {
-      return parser.parseFrom(input);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseWithIOException(Parser<M> parser, InputStream input,
-      ExtensionRegistryLite extensions) throws IOException {
-    try {
-      return parser.parseFrom(input, extensions);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseWithIOException(Parser<M> parser,
-      CodedInputStream input) throws IOException {
-    try {
-      return parser.parseFrom(input);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseWithIOException(Parser<M> parser,
-      CodedInputStream input, ExtensionRegistryLite extensions) throws IOException {
-    try {
-      return parser.parseFrom(input, extensions);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseDelimitedWithIOException(Parser<M> parser,
-      InputStream input) throws IOException {
-    try {
-      return parser.parseDelimitedFrom(input);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseDelimitedWithIOException(Parser<M> parser,
-      InputStream input, ExtensionRegistryLite extensions) throws IOException {
-    try {
-      return parser.parseDelimitedFrom(input, extensions);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  @Override
-  public void writeTo(final CodedOutputStream output) throws IOException {
-    MessageReflection.writeMessageTo(this, getAllFieldsRaw(), output, false);
-  }
-
-  @Override
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) {
-      return size;
-    }
-
-    memoizedSize = MessageReflection.getSerializedSize(
-        this, getAllFieldsRaw());
-    return memoizedSize;
-  }
-
-
-
-  /**
-   * Used by parsing constructors in generated classes.
-   */
-  protected void makeExtensionsImmutable() {
-    // Noop for messages without extensions.
-  }
-
-  /**
-   * TODO(xiaofeng): remove this after b/29368482 is fixed. We need to move this
-   * interface to AbstractMessage in order to versioning GeneratedMessageV3 but
-   * this move breaks binary compatibility for AppEngine. After AppEngine is
-   * fixed we can exlude this from google3.
-   */
-  protected interface BuilderParent extends AbstractMessage.BuilderParent {}
-
-  /**
-   * TODO(xiaofeng): remove this together with GeneratedMessageV3.BuilderParent.
-   */
-  protected abstract Message.Builder newBuilderForType(BuilderParent parent);
-
-  @Override
-  protected Message.Builder newBuilderForType(final AbstractMessage.BuilderParent parent) {
-    return newBuilderForType(new BuilderParent() {
-      @Override
-      public void markDirty() {
-        parent.markDirty();
-      }
-    });
-  }
-
-
-  @SuppressWarnings("unchecked")
-  public abstract static class Builder <BuilderType extends Builder<BuilderType>>
-      extends AbstractMessage.Builder<BuilderType> {
-
-    private BuilderParent builderParent;
-
-    private BuilderParentImpl meAsParent;
-
-    // Indicates that we've built a message and so we are now obligated
-    // to dispatch dirty invalidations. See GeneratedMessageV3.BuilderListener.
-    private boolean isClean;
-
-    private UnknownFieldSet unknownFields =
-        UnknownFieldSet.getDefaultInstance();
-
-    protected Builder() {
-      this(null);
-    }
-
-    protected Builder(BuilderParent builderParent) {
-      this.builderParent = builderParent;
-    }
-
-    @Override
-    void dispose() {
-      builderParent = null;
-    }
-
-    /**
-     * Called by the subclass when a message is built.
-     */
-    protected void onBuilt() {
-      if (builderParent != null) {
-        markClean();
-      }
-    }
-
-    /**
-     * Called by the subclass or a builder to notify us that a message was
-     * built and may be cached and therefore invalidations are needed.
-     */
-    @Override
-    protected void markClean() {
-      this.isClean = true;
-    }
-
-    /**
-     * Gets whether invalidations are needed
-     *
-     * @return whether invalidations are needed
-     */
-    protected boolean isClean() {
-      return isClean;
-    }
-
-    @Override
-    public BuilderType clone() {
-      BuilderType builder =
-          (BuilderType) getDefaultInstanceForType().newBuilderForType();
-      builder.mergeFrom(buildPartial());
-      return builder;
-    }
-
-    /**
-     * Called by the initialization and clear code paths to allow subclasses to
-     * reset any of their builtin fields back to the initial values.
-     */
-    @Override
-    public BuilderType clear() {
-      unknownFields = UnknownFieldSet.getDefaultInstance();
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /**
-     * Get the FieldAccessorTable for this type.  We can't have the message
-     * class pass this in to the constructor because of bootstrapping trouble
-     * with DescriptorProtos.
-     */
-    protected abstract FieldAccessorTable internalGetFieldAccessorTable();
-
-    @Override
-    public Descriptor getDescriptorForType() {
-      return internalGetFieldAccessorTable().descriptor;
-    }
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFields() {
-      return Collections.unmodifiableMap(getAllFieldsMutable());
-    }
-
-    /** Internal helper which returns a mutable map. */
-    private Map<FieldDescriptor, Object> getAllFieldsMutable() {
-      final TreeMap<FieldDescriptor, Object> result =
-        new TreeMap<FieldDescriptor, Object>();
-      final Descriptor descriptor = internalGetFieldAccessorTable().descriptor;
-      final List<FieldDescriptor> fields = descriptor.getFields();
-
-      for (int i = 0; i < fields.size(); i++) {
-        FieldDescriptor field = fields.get(i);
-        final OneofDescriptor oneofDescriptor = field.getContainingOneof();
-
-        /*
-         * If the field is part of a Oneof, then at maximum one field in the Oneof is set
-         * and it is not repeated. There is no need to iterate through the others.
-         */
-        if (oneofDescriptor != null) {
-          // Skip other fields in the Oneof we know are not set
-          i += oneofDescriptor.getFieldCount() - 1;
-          if (!hasOneof(oneofDescriptor)) {
-            // If no field is set in the Oneof, skip all the fields in the Oneof
-            continue;
-          }
-          // Get the pointer to the only field which is set in the Oneof
-          field = getOneofFieldDescriptor(oneofDescriptor);
-        } else {
-          // If we are not in a Oneof, we need to check if the field is set and if it is repeated
-          if (field.isRepeated()) {
-            final List<?> value = (List<?>) getField(field);
-            if (!value.isEmpty()) {
-              result.put(field, value);
-            }
-            continue;
-          }
-          if (!hasField(field)) {
-            continue;
-          }
-        }
-        // Add the field to the map
-        result.put(field, getField(field));
-      }
-      return result;
-    }
-
-    @Override
-    public Message.Builder newBuilderForField(final FieldDescriptor field) {
-      return internalGetFieldAccessorTable().getField(field).newBuilder();
-    }
-
-    @Override
-    public Message.Builder getFieldBuilder(final FieldDescriptor field) {
-      return internalGetFieldAccessorTable().getField(field).getBuilder(this);
-    }
-
-    @Override
-    public Message.Builder getRepeatedFieldBuilder(final FieldDescriptor field, int index) {
-      return internalGetFieldAccessorTable().getField(field).getRepeatedBuilder(
-          this, index);
-    }
-
-    @Override
-    public boolean hasOneof(final OneofDescriptor oneof) {
-      return internalGetFieldAccessorTable().getOneof(oneof).has(this);
-    }
-
-    @Override
-    public FieldDescriptor getOneofFieldDescriptor(final OneofDescriptor oneof) {
-      return internalGetFieldAccessorTable().getOneof(oneof).get(this);
-    }
-
-    @Override
-    public boolean hasField(final FieldDescriptor field) {
-      return internalGetFieldAccessorTable().getField(field).has(this);
-    }
-
-    @Override
-    public Object getField(final FieldDescriptor field) {
-      Object object = internalGetFieldAccessorTable().getField(field).get(this);
-      if (field.isRepeated()) {
-        // The underlying list object is still modifiable at this point.
-        // Make sure not to expose the modifiable list to the caller.
-        return Collections.unmodifiableList((List) object);
-      } else {
-        return object;
-      }
-    }
-
-    @Override
-    public BuilderType setField(final FieldDescriptor field, final Object value) {
-      internalGetFieldAccessorTable().getField(field).set(this, value);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType clearField(final FieldDescriptor field) {
-      internalGetFieldAccessorTable().getField(field).clear(this);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType clearOneof(final OneofDescriptor oneof) {
-      internalGetFieldAccessorTable().getOneof(oneof).clear(this);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public int getRepeatedFieldCount(final FieldDescriptor field) {
-      return internalGetFieldAccessorTable().getField(field)
-          .getRepeatedCount(this);
-    }
-
-    @Override
-    public Object getRepeatedField(final FieldDescriptor field, final int index) {
-      return internalGetFieldAccessorTable().getField(field)
-          .getRepeated(this, index);
-    }
-
-    @Override
-    public BuilderType setRepeatedField(
-        final FieldDescriptor field, final int index, final Object value) {
-      internalGetFieldAccessorTable().getField(field)
-        .setRepeated(this, index, value);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType addRepeatedField(final FieldDescriptor field, final Object value) {
-      internalGetFieldAccessorTable().getField(field).addRepeated(this, value);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType setUnknownFields(final UnknownFieldSet unknownFields) {
-      this.unknownFields = unknownFields;
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType mergeUnknownFields(
-        final UnknownFieldSet unknownFields) {
-      this.unknownFields =
-        UnknownFieldSet.newBuilder(this.unknownFields)
-                       .mergeFrom(unknownFields)
-                       .build();
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    @Override
-    public boolean isInitialized() {
-      for (final FieldDescriptor field : getDescriptorForType().getFields()) {
-        // Check that all required fields are present.
-        if (field.isRequired()) {
-          if (!hasField(field)) {
-            return false;
-          }
-        }
-        // Check that embedded messages are initialized.
-        if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-          if (field.isRepeated()) {
-            @SuppressWarnings("unchecked") final
-            List<Message> messageList = (List<Message>) getField(field);
-            for (final Message element : messageList) {
-              if (!element.isInitialized()) {
-                return false;
-              }
-            }
-          } else {
-            if (hasField(field) &&
-                !((Message) getField(field)).isInitialized()) {
-              return false;
-            }
-          }
-        }
-      }
-      return true;
-    }
-
-    @Override
-    public final UnknownFieldSet getUnknownFields() {
-      return unknownFields;
-    }
-
-    /**
-     * Called by subclasses to parse an unknown field.
-     * @return {@code true} unless the tag is an end-group tag.
-     */
-    protected boolean parseUnknownField(
-        final CodedInputStream input,
-        final UnknownFieldSet.Builder unknownFields,
-        final ExtensionRegistryLite extensionRegistry,
-        final int tag) throws IOException {
-      return unknownFields.mergeFieldFrom(tag, input);
-    }
-
-    /**
-     * Implementation of {@link BuilderParent} for giving to our children. This
-     * small inner class makes it so we don't publicly expose the BuilderParent
-     * methods.
-     */
-    private class BuilderParentImpl implements BuilderParent {
-
-      @Override
-      public void markDirty() {
-        onChanged();
-      }
-    }
-
-    /**
-     * Gets the {@link BuilderParent} for giving to our children.
-     * @return The builder parent for our children.
-     */
-    protected BuilderParent getParentForChildren() {
-      if (meAsParent == null) {
-        meAsParent = new BuilderParentImpl();
-      }
-      return meAsParent;
-    }
-
-    /**
-     * Called when a the builder or one of its nested children has changed
-     * and any parent should be notified of its invalidation.
-     */
-    protected final void onChanged() {
-      if (isClean && builderParent != null) {
-        builderParent.markDirty();
-
-        // Don't keep dispatching invalidations until build is called again.
-        isClean = false;
-      }
-    }
-
-    /**
-     * Gets the map field with the given field number. This method should be
-     * overridden in the generated message class if the message contains map
-     * fields.
-     *
-     * Unlike other field types, reflection support for map fields can't be
-     * implemented based on generated public API because we need to access a
-     * map field as a list in reflection API but the generated API only allows
-     * us to access it as a map. This method returns the underlying map field
-     * directly and thus enables us to access the map field as a list.
-     */
-    @SuppressWarnings({"unused", "rawtypes"})
-    protected MapField internalGetMapField(int fieldNumber) {
-      // Note that we can't use descriptor names here because this method will
-      // be called when descriptor is being initialized.
-      throw new RuntimeException(
-          "No map fields found in " + getClass().getName());
-    }
-
-    /** Like {@link #internalGetMapField} but return a mutable version. */
-    @SuppressWarnings({"unused", "rawtypes"})
-    protected MapField internalGetMutableMapField(int fieldNumber) {
-      // Note that we can't use descriptor names here because this method will
-      // be called when descriptor is being initialized.
-      throw new RuntimeException(
-          "No map fields found in " + getClass().getName());
-    }
-  }
-
-  // =================================================================
-  // Extensions-related stuff
-
-  public interface ExtendableMessageOrBuilder<
-      MessageType extends ExtendableMessage> extends MessageOrBuilder {
-    // Re-define for return type covariance.
-    @Override
-    Message getDefaultInstanceForType();
-
-    /** Check if a singular extension is present. */
-    <Type> boolean hasExtension(
-        ExtensionLite<MessageType, Type> extension);
-
-    /** Get the number of elements in a repeated extension. */
-    <Type> int getExtensionCount(
-        ExtensionLite<MessageType, List<Type>> extension);
-
-    /** Get the value of an extension. */
-    <Type> Type getExtension(
-        ExtensionLite<MessageType, Type> extension);
-
-    /** Get one element of a repeated extension. */
-    <Type> Type getExtension(
-        ExtensionLite<MessageType, List<Type>> extension,
-        int index);
-
-    /** Check if a singular extension is present. */
-    <Type> boolean hasExtension(
-        Extension<MessageType, Type> extension);
-    /** Check if a singular extension is present. */
-    <Type> boolean hasExtension(
-        GeneratedExtension<MessageType, Type> extension);
-    /** Get the number of elements in a repeated extension. */
-    <Type> int getExtensionCount(
-        Extension<MessageType, List<Type>> extension);
-    /** Get the number of elements in a repeated extension. */
-    <Type> int getExtensionCount(
-        GeneratedExtension<MessageType, List<Type>> extension);
-    /** Get the value of an extension. */
-    <Type> Type getExtension(
-        Extension<MessageType, Type> extension);
-    /** Get the value of an extension. */
-    <Type> Type getExtension(
-        GeneratedExtension<MessageType, Type> extension);
-    /** Get one element of a repeated extension. */
-    <Type> Type getExtension(
-        Extension<MessageType, List<Type>> extension,
-        int index);
-    /** Get one element of a repeated extension. */
-    <Type> Type getExtension(
-        GeneratedExtension<MessageType, List<Type>> extension,
-        int index);
-  }
-
-  /**
-   * Generated message classes for message types that contain extension ranges
-   * subclass this.
-   *
-   * <p>This class implements type-safe accessors for extensions.  They
-   * implement all the same operations that you can do with normal fields --
-   * e.g. "has", "get", and "getCount" -- but for extensions.  The extensions
-   * are identified using instances of the class {@link GeneratedExtension};
-   * the protocol compiler generates a static instance of this class for every
-   * extension in its input.  Through the magic of generics, all is made
-   * type-safe.
-   *
-   * <p>For example, imagine you have the {@code .proto} file:
-   *
-   * <pre>
-   * option java_class = "MyProto";
-   *
-   * message Foo {
-   *   extensions 1000 to max;
-   * }
-   *
-   * extend Foo {
-   *   optional int32 bar;
-   * }
-   * </pre>
-   *
-   * <p>Then you might write code like:
-   *
-   * <pre>
-   * MyProto.Foo foo = getFoo();
-   * int i = foo.getExtension(MyProto.bar);
-   * </pre>
-   *
-   * <p>See also {@link ExtendableBuilder}.
-   */
-  public abstract static class ExtendableMessage<
-        MessageType extends ExtendableMessage>
-      extends GeneratedMessageV3
-      implements ExtendableMessageOrBuilder<MessageType> {
-
-    private static final long serialVersionUID = 1L;
-
-    private final FieldSet<FieldDescriptor> extensions;
-
-    protected ExtendableMessage() {
-      this.extensions = FieldSet.newFieldSet();
-    }
-
-    protected ExtendableMessage(
-        ExtendableBuilder<MessageType, ?> builder) {
-      super(builder);
-      this.extensions = builder.buildExtensions();
-    }
-
-    private void verifyExtensionContainingType(
-        final Extension<MessageType, ?> extension) {
-      if (extension.getDescriptor().getContainingType() !=
-          getDescriptorForType()) {
-        // This can only happen if someone uses unchecked operations.
-        throw new IllegalArgumentException(
-          "Extension is for type \"" +
-          extension.getDescriptor().getContainingType().getFullName() +
-          "\" which does not match message type \"" +
-          getDescriptorForType().getFullName() + "\".");
-      }
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final ExtensionLite<MessageType, Type> extensionLite) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      return extensions.hasField(extension.getDescriptor());
-    }
-
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final ExtensionLite<MessageType, List<Type>> extensionLite) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      return extensions.getRepeatedFieldCount(descriptor);
-    }
-
-    /** Get the value of an extension. */
-    @Override
-    @SuppressWarnings("unchecked")
-    public final <Type> Type getExtension(final ExtensionLite<MessageType, Type> extensionLite) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      FieldDescriptor descriptor = extension.getDescriptor();
-      final Object value = extensions.getField(descriptor);
-      if (value == null) {
-        if (descriptor.isRepeated()) {
-          return (Type) Collections.emptyList();
-        } else if (descriptor.getJavaType() ==
-                   FieldDescriptor.JavaType.MESSAGE) {
-          return (Type) extension.getMessageDefaultInstance();
-        } else {
-          return (Type) extension.fromReflectionType(
-              descriptor.getDefaultValue());
-        }
-      } else {
-        return (Type) extension.fromReflectionType(value);
-      }
-    }
-
-    /** Get one element of a repeated extension. */
-    @Override
-    @SuppressWarnings("unchecked")
-    public final <Type> Type getExtension(
-        final ExtensionLite<MessageType, List<Type>> extensionLite, final int index) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      FieldDescriptor descriptor = extension.getDescriptor();
-      return (Type) extension.singularFromReflectionType(
-          extensions.getRepeatedField(descriptor, index));
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final Extension<MessageType, Type> extension) {
-      return hasExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(
-        final GeneratedExtension<MessageType, Type> extension) {
-      return hasExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final Extension<MessageType, List<Type>> extension) {
-      return getExtensionCount((ExtensionLite<MessageType, List<Type>>) extension);
-    }
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final GeneratedExtension<MessageType, List<Type>> extension) {
-      return getExtensionCount((ExtensionLite<MessageType, List<Type>>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(final Extension<MessageType, Type> extension) {
-      return getExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final GeneratedExtension<MessageType, Type> extension) {
-      return getExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get one element of a repeated extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final Extension<MessageType, List<Type>> extension, final int index) {
-      return getExtension((ExtensionLite<MessageType, List<Type>>) extension, index);
-    }
-    /** Get one element of a repeated extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final GeneratedExtension<MessageType, List<Type>> extension, final int index) {
-      return getExtension((ExtensionLite<MessageType, List<Type>>) extension, index);
-    }
-
-    /** Called by subclasses to check if all extensions are initialized. */
-    protected boolean extensionsAreInitialized() {
-      return extensions.isInitialized();
-    }
-
-    @Override
-    public boolean isInitialized() {
-      return super.isInitialized() && extensionsAreInitialized();
-    }
-
-    @Override
-    protected boolean parseUnknownField(
-        CodedInputStream input,
-        UnknownFieldSet.Builder unknownFields,
-        ExtensionRegistryLite extensionRegistry,
-        int tag) throws IOException {
-      return MessageReflection.mergeFieldFrom(
-          input, unknownFields, extensionRegistry, getDescriptorForType(),
-          new MessageReflection.ExtensionAdapter(extensions), tag);
-    }
-
-
-    /**
-     * Used by parsing constructors in generated classes.
-     */
-    @Override
-    protected void makeExtensionsImmutable() {
-      extensions.makeImmutable();
-    }
-
-    /**
-     * Used by subclasses to serialize extensions.  Extension ranges may be
-     * interleaved with field numbers, but we must write them in canonical
-     * (sorted by field number) order.  ExtensionWriter helps us write
-     * individual ranges of extensions at once.
-     */
-    protected class ExtensionWriter {
-      // Imagine how much simpler this code would be if Java iterators had
-      // a way to get the next element without advancing the iterator.
-
-      private final Iterator<Map.Entry<FieldDescriptor, Object>> iter =
-        extensions.iterator();
-      private Map.Entry<FieldDescriptor, Object> next;
-      private final boolean messageSetWireFormat;
-
-      private ExtensionWriter(final boolean messageSetWireFormat) {
-        if (iter.hasNext()) {
-          next = iter.next();
-        }
-        this.messageSetWireFormat = messageSetWireFormat;
-      }
-
-      public void writeUntil(final int end, final CodedOutputStream output)
-                             throws IOException {
-        while (next != null && next.getKey().getNumber() < end) {
-          FieldDescriptor descriptor = next.getKey();
-          if (messageSetWireFormat && descriptor.getLiteJavaType() ==
-                  WireFormat.JavaType.MESSAGE &&
-              !descriptor.isRepeated()) {
-            if (next instanceof LazyField.LazyEntry<?>) {
-              output.writeRawMessageSetExtension(descriptor.getNumber(),
-                  ((LazyField.LazyEntry<?>) next).getField().toByteString());
-            } else {
-              output.writeMessageSetExtension(descriptor.getNumber(),
-                                              (Message) next.getValue());
-            }
-          } else {
-            // TODO(xiangl): Taken care of following code, it may cause
-            // problem when we use LazyField for normal fields/extensions.
-            // Due to the optional field can be duplicated at the end of
-            // serialized bytes, which will make the serialized size change
-            // after lazy field parsed. So when we use LazyField globally,
-            // we need to change the following write method to write cached
-            // bytes directly rather than write the parsed message.
-            FieldSet.writeField(descriptor, next.getValue(), output);
-          }
-          if (iter.hasNext()) {
-            next = iter.next();
-          } else {
-            next = null;
-          }
-        }
-      }
-    }
-
-    protected ExtensionWriter newExtensionWriter() {
-      return new ExtensionWriter(false);
-    }
-    protected ExtensionWriter newMessageSetExtensionWriter() {
-      return new ExtensionWriter(true);
-    }
-
-    /** Called by subclasses to compute the size of extensions. */
-    protected int extensionsSerializedSize() {
-      return extensions.getSerializedSize();
-    }
-    protected int extensionsSerializedSizeAsMessageSet() {
-      return extensions.getMessageSetSerializedSize();
-    }
-
-    // ---------------------------------------------------------------
-    // Reflection
-
-    protected Map<FieldDescriptor, Object> getExtensionFields() {
-      return extensions.getAllFields();
-    }
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFields() {
-      final Map<FieldDescriptor, Object> result =
-          super.getAllFieldsMutable(/* getBytesForString = */ false);
-      result.putAll(getExtensionFields());
-      return Collections.unmodifiableMap(result);
-    }
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFieldsRaw() {
-      final Map<FieldDescriptor, Object> result =
-          super.getAllFieldsMutable(/* getBytesForString = */ false);
-      result.putAll(getExtensionFields());
-      return Collections.unmodifiableMap(result);
-    }
-
-    @Override
-    public boolean hasField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.hasField(field);
-      } else {
-        return super.hasField(field);
-      }
-    }
-
-    @Override
-    public Object getField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        final Object value = extensions.getField(field);
-        if (value == null) {
-          if (field.isRepeated()) {
-            return Collections.emptyList();
-          } else if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-            // Lacking an ExtensionRegistry, we have no way to determine the
-            // extension's real type, so we return a DynamicMessage.
-            return DynamicMessage.getDefaultInstance(field.getMessageType());
-          } else {
-            return field.getDefaultValue();
-          }
-        } else {
-          return value;
-        }
-      } else {
-        return super.getField(field);
-      }
-    }
-
-    @Override
-    public int getRepeatedFieldCount(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.getRepeatedFieldCount(field);
-      } else {
-        return super.getRepeatedFieldCount(field);
-      }
-    }
-
-    @Override
-    public Object getRepeatedField(final FieldDescriptor field,
-                                   final int index) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.getRepeatedField(field, index);
-      } else {
-        return super.getRepeatedField(field, index);
-      }
-    }
-
-    private void verifyContainingType(final FieldDescriptor field) {
-      if (field.getContainingType() != getDescriptorForType()) {
-        throw new IllegalArgumentException(
-          "FieldDescriptor does not match message type.");
-      }
-    }
-  }
-
-  /**
-   * Generated message builders for message types that contain extension ranges
-   * subclass this.
-   *
-   * <p>This class implements type-safe accessors for extensions.  They
-   * implement all the same operations that you can do with normal fields --
-   * e.g. "get", "set", and "add" -- but for extensions.  The extensions are
-   * identified using instances of the class {@link GeneratedExtension}; the
-   * protocol compiler generates a static instance of this class for every
-   * extension in its input.  Through the magic of generics, all is made
-   * type-safe.
-   *
-   * <p>For example, imagine you have the {@code .proto} file:
-   *
-   * <pre>
-   * option java_class = "MyProto";
-   *
-   * message Foo {
-   *   extensions 1000 to max;
-   * }
-   *
-   * extend Foo {
-   *   optional int32 bar;
-   * }
-   * </pre>
-   *
-   * <p>Then you might write code like:
-   *
-   * <pre>
-   * MyProto.Foo foo =
-   *   MyProto.Foo.newBuilder()
-   *     .setExtension(MyProto.bar, 123)
-   *     .build();
-   * </pre>
-   *
-   * <p>See also {@link ExtendableMessage}.
-   */
-  @SuppressWarnings("unchecked")
-  public abstract static class ExtendableBuilder<
-        MessageType extends ExtendableMessage,
-        BuilderType extends ExtendableBuilder<MessageType, BuilderType>>
-      extends Builder<BuilderType>
-      implements ExtendableMessageOrBuilder<MessageType> {
-
-    private FieldSet<FieldDescriptor> extensions = FieldSet.emptySet();
-
-    protected ExtendableBuilder() {}
-
-    protected ExtendableBuilder(
-        BuilderParent parent) {
-      super(parent);
-    }
-
-    // For immutable message conversion.
-    void internalSetExtensionSet(FieldSet<FieldDescriptor> extensions) {
-      this.extensions = extensions;
-    }
-
-    @Override
-    public BuilderType clear() {
-      extensions = FieldSet.emptySet();
-      return super.clear();
-    }
-
-    private void ensureExtensionsIsMutable() {
-      if (extensions.isImmutable()) {
-        extensions = extensions.clone();
-      }
-    }
-
-    private void verifyExtensionContainingType(
-        final Extension<MessageType, ?> extension) {
-      if (extension.getDescriptor().getContainingType() !=
-          getDescriptorForType()) {
-        // This can only happen if someone uses unchecked operations.
-        throw new IllegalArgumentException(
-          "Extension is for type \"" +
-          extension.getDescriptor().getContainingType().getFullName() +
-          "\" which does not match message type \"" +
-          getDescriptorForType().getFullName() + "\".");
-      }
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final ExtensionLite<MessageType, Type> extensionLite) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      return extensions.hasField(extension.getDescriptor());
-    }
-
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final ExtensionLite<MessageType, List<Type>> extensionLite) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      return extensions.getRepeatedFieldCount(descriptor);
-    }
-
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(final ExtensionLite<MessageType, Type> extensionLite) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      FieldDescriptor descriptor = extension.getDescriptor();
-      final Object value = extensions.getField(descriptor);
-      if (value == null) {
-        if (descriptor.isRepeated()) {
-          return (Type) Collections.emptyList();
-        } else if (descriptor.getJavaType() ==
-                   FieldDescriptor.JavaType.MESSAGE) {
-          return (Type) extension.getMessageDefaultInstance();
-        } else {
-          return (Type) extension.fromReflectionType(
-              descriptor.getDefaultValue());
-        }
-      } else {
-        return (Type) extension.fromReflectionType(value);
-      }
-    }
-
-    /** Get one element of a repeated extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final ExtensionLite<MessageType, List<Type>> extensionLite, final int index) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      FieldDescriptor descriptor = extension.getDescriptor();
-      return (Type) extension.singularFromReflectionType(
-          extensions.getRepeatedField(descriptor, index));
-    }
-
-    /** Set the value of an extension. */
-    public final <Type> BuilderType setExtension(
-        final ExtensionLite<MessageType, Type> extensionLite,
-        final Type value) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      ensureExtensionsIsMutable();
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      extensions.setField(descriptor, extension.toReflectionType(value));
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /** Set the value of one element of a repeated extension. */
-    public final <Type> BuilderType setExtension(
-        final ExtensionLite<MessageType, List<Type>> extensionLite,
-        final int index, final Type value) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      ensureExtensionsIsMutable();
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      extensions.setRepeatedField(
-        descriptor, index,
-        extension.singularToReflectionType(value));
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /** Append a value to a repeated extension. */
-    public final <Type> BuilderType addExtension(
-        final ExtensionLite<MessageType, List<Type>> extensionLite,
-        final Type value) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      ensureExtensionsIsMutable();
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      extensions.addRepeatedField(
-          descriptor, extension.singularToReflectionType(value));
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /** Clear an extension. */
-    public final <Type> BuilderType clearExtension(
-        final ExtensionLite<MessageType, ?> extensionLite) {
-      Extension<MessageType, ?> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      ensureExtensionsIsMutable();
-      extensions.clearField(extension.getDescriptor());
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final Extension<MessageType, Type> extension) {
-      return hasExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(
-        final GeneratedExtension<MessageType, Type> extension) {
-      return hasExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final Extension<MessageType, List<Type>> extension) {
-      return getExtensionCount((ExtensionLite<MessageType, List<Type>>) extension);
-    }
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final GeneratedExtension<MessageType, List<Type>> extension) {
-      return getExtensionCount((ExtensionLite<MessageType, List<Type>>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(final Extension<MessageType, Type> extension) {
-      return getExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final GeneratedExtension<MessageType, Type> extension) {
-      return getExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final Extension<MessageType, List<Type>> extension, final int index) {
-      return getExtension((ExtensionLite<MessageType, List<Type>>) extension, index);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final GeneratedExtension<MessageType, List<Type>> extension, final int index) {
-      return getExtension((ExtensionLite<MessageType, List<Type>>) extension, index);
-    }
-    /** Set the value of an extension. */
-    public final <Type> BuilderType setExtension(
-        final Extension<MessageType, Type> extension, final Type value) {
-      return setExtension((ExtensionLite<MessageType, Type>) extension, value);
-    }
-    /** Set the value of an extension. */
-    public <Type> BuilderType setExtension(
-        final GeneratedExtension<MessageType, Type> extension, final Type value) {
-      return setExtension((ExtensionLite<MessageType, Type>) extension, value);
-    }
-    /** Set the value of one element of a repeated extension. */
-    public final <Type> BuilderType setExtension(
-        final Extension<MessageType, List<Type>> extension,
-        final int index, final Type value) {
-      return setExtension((ExtensionLite<MessageType, List<Type>>) extension, index, value);
-    }
-    /** Set the value of one element of a repeated extension. */
-    public <Type> BuilderType setExtension(
-        final GeneratedExtension<MessageType, List<Type>> extension,
-        final int index, final Type value) {
-      return setExtension((ExtensionLite<MessageType, List<Type>>) extension, index, value);
-    }
-    /** Append a value to a repeated extension. */
-    public final <Type> BuilderType addExtension(
-        final Extension<MessageType, List<Type>> extension, final Type value) {
-      return addExtension((ExtensionLite<MessageType, List<Type>>) extension, value);
-    }
-    /** Append a value to a repeated extension. */
-    public <Type> BuilderType addExtension(
-        final GeneratedExtension<MessageType, List<Type>> extension, final Type value) {
-      return addExtension((ExtensionLite<MessageType, List<Type>>) extension, value);
-    }
-    /** Clear an extension. */
-    public final <Type> BuilderType clearExtension(
-        final Extension<MessageType, ?> extension) {
-      return clearExtension((ExtensionLite<MessageType, ?>) extension);
-    }
-    /** Clear an extension. */
-    public <Type> BuilderType clearExtension(
-        final GeneratedExtension<MessageType, ?> extension) {
-      return clearExtension((ExtensionLite<MessageType, ?>) extension);
-    }
-
-    /** Called by subclasses to check if all extensions are initialized. */
-    protected boolean extensionsAreInitialized() {
-      return extensions.isInitialized();
-    }
-
-    /**
-     * Called by the build code path to create a copy of the extensions for
-     * building the message.
-     */
-    private FieldSet<FieldDescriptor> buildExtensions() {
-      extensions.makeImmutable();
-      return extensions;
-    }
-
-    @Override
-    public boolean isInitialized() {
-      return super.isInitialized() && extensionsAreInitialized();
-    }
-
-    /**
-     * Called by subclasses to parse an unknown field or an extension.
-     * @return {@code true} unless the tag is an end-group tag.
-     */
-    @Override
-    protected boolean parseUnknownField(
-        final CodedInputStream input,
-        final UnknownFieldSet.Builder unknownFields,
-        final ExtensionRegistryLite extensionRegistry,
-        final int tag) throws IOException {
-      return MessageReflection.mergeFieldFrom(
-          input, unknownFields, extensionRegistry, getDescriptorForType(),
-          new MessageReflection.BuilderAdapter(this), tag);
-    }
-
-    // ---------------------------------------------------------------
-    // Reflection
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFields() {
-      final Map<FieldDescriptor, Object> result = super.getAllFieldsMutable();
-      result.putAll(extensions.getAllFields());
-      return Collections.unmodifiableMap(result);
-    }
-
-    @Override
-    public Object getField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        final Object value = extensions.getField(field);
-        if (value == null) {
-          if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-            // Lacking an ExtensionRegistry, we have no way to determine the
-            // extension's real type, so we return a DynamicMessage.
-            return DynamicMessage.getDefaultInstance(field.getMessageType());
-          } else {
-            return field.getDefaultValue();
-          }
-        } else {
-          return value;
-        }
-      } else {
-        return super.getField(field);
-      }
-    }
-
-    @Override
-    public int getRepeatedFieldCount(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.getRepeatedFieldCount(field);
-      } else {
-        return super.getRepeatedFieldCount(field);
-      }
-    }
-
-    @Override
-    public Object getRepeatedField(final FieldDescriptor field,
-                                   final int index) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.getRepeatedField(field, index);
-      } else {
-        return super.getRepeatedField(field, index);
-      }
-    }
-
-    @Override
-    public boolean hasField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.hasField(field);
-      } else {
-        return super.hasField(field);
-      }
-    }
-
-    @Override
-    public BuilderType setField(final FieldDescriptor field,
-                                final Object value) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        ensureExtensionsIsMutable();
-        extensions.setField(field, value);
-        onChanged();
-        return (BuilderType) this;
-      } else {
-        return super.setField(field, value);
-      }
-    }
-
-    @Override
-    public BuilderType clearField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        ensureExtensionsIsMutable();
-        extensions.clearField(field);
-        onChanged();
-        return (BuilderType) this;
-      } else {
-        return super.clearField(field);
-      }
-    }
-
-    @Override
-    public BuilderType setRepeatedField(final FieldDescriptor field,
-                                        final int index, final Object value) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        ensureExtensionsIsMutable();
-        extensions.setRepeatedField(field, index, value);
-        onChanged();
-        return (BuilderType) this;
-      } else {
-        return super.setRepeatedField(field, index, value);
-      }
-    }
-
-    @Override
-    public BuilderType addRepeatedField(final FieldDescriptor field,
-                                        final Object value) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        ensureExtensionsIsMutable();
-        extensions.addRepeatedField(field, value);
-        onChanged();
-        return (BuilderType) this;
-      } else {
-        return super.addRepeatedField(field, value);
-      }
-    }
-
-    protected final void mergeExtensionFields(final ExtendableMessage other) {
-      ensureExtensionsIsMutable();
-      extensions.mergeFrom(other.extensions);
-      onChanged();
-    }
-
-    private void verifyContainingType(final FieldDescriptor field) {
-      if (field.getContainingType() != getDescriptorForType()) {
-        throw new IllegalArgumentException(
-          "FieldDescriptor does not match message type.");
-      }
-    }
-  }
-
-  // -----------------------------------------------------------------
-
-  /**
-   * Gets the descriptor for an extension. The implementation depends on whether
-   * the extension is scoped in the top level of a file or scoped in a Message.
-   */
-  static interface ExtensionDescriptorRetriever {
-    FieldDescriptor getDescriptor();
-  }
-
-
-  // =================================================================
-
-  /** Calls Class.getMethod and throws a RuntimeException if it fails. */
-  @SuppressWarnings("unchecked")
-  private static Method getMethodOrDie(
-      final Class clazz, final String name, final Class... params) {
-    try {
-      return clazz.getMethod(name, params);
-    } catch (NoSuchMethodException e) {
-      throw new RuntimeException(
-        "Generated message class \"" + clazz.getName() +
-        "\" missing method \"" + name + "\".", e);
-    }
-  }
-
-  /** Calls invoke and throws a RuntimeException if it fails. */
-  private static Object invokeOrDie(
-      final Method method, final Object object, final Object... params) {
-    try {
-      return method.invoke(object, params);
-    } catch (IllegalAccessException e) {
-      throw new RuntimeException(
-        "Couldn't use Java reflection to implement protocol message " +
-        "reflection.", e);
-    } catch (InvocationTargetException e) {
-      final Throwable cause = e.getCause();
-      if (cause instanceof RuntimeException) {
-        throw (RuntimeException) cause;
-      } else if (cause instanceof Error) {
-        throw (Error) cause;
-      } else {
-        throw new RuntimeException(
-          "Unexpected exception thrown by generated accessor method.", cause);
-      }
-    }
-  }
-
-  /**
-   * Gets the map field with the given field number. This method should be
-   * overridden in the generated message class if the message contains map
-   * fields.
-   *
-   * Unlike other field types, reflection support for map fields can't be
-   * implemented based on generated public API because we need to access a
-   * map field as a list in reflection API but the generated API only allows
-   * us to access it as a map. This method returns the underlying map field
-   * directly and thus enables us to access the map field as a list.
-   */
-  @SuppressWarnings({"rawtypes", "unused"})
-  protected MapField internalGetMapField(int fieldNumber) {
-    // Note that we can't use descriptor names here because this method will
-    // be called when descriptor is being initialized.
-    throw new RuntimeException(
-        "No map fields found in " + getClass().getName());
-  }
-
-  /**
-   * Users should ignore this class.  This class provides the implementation
-   * with access to the fields of a message object using Java reflection.
-   */
-  public static final class FieldAccessorTable {
-
-    /**
-     * Construct a FieldAccessorTable for a particular message class.  Only
-     * one FieldAccessorTable should ever be constructed per class.
-     *
-     * @param descriptor     The type's descriptor.
-     * @param camelCaseNames The camelcase names of all fields in the message.
-     *                       These are used to derive the accessor method names.
-     * @param messageClass   The message type.
-     * @param builderClass   The builder type.
-     */
-    public FieldAccessorTable(
-        final Descriptor descriptor,
-        final String[] camelCaseNames,
-        final Class<? extends GeneratedMessageV3> messageClass,
-        final Class<? extends Builder> builderClass) {
-      this(descriptor, camelCaseNames);
-      ensureFieldAccessorsInitialized(messageClass, builderClass);
-    }
-
-    /**
-     * Construct a FieldAccessorTable for a particular message class without
-     * initializing FieldAccessors.
-     */
-    public FieldAccessorTable(
-        final Descriptor descriptor,
-        final String[] camelCaseNames) {
-      this.descriptor = descriptor;
-      this.camelCaseNames = camelCaseNames;
-      fields = new FieldAccessor[descriptor.getFields().size()];
-      oneofs = new OneofAccessor[descriptor.getOneofs().size()];
-      initialized = false;
-    }
-
-    /**
-     * Ensures the field accessors are initialized. This method is thread-safe.
-     *
-     * @param messageClass   The message type.
-     * @param builderClass   The builder type.
-     * @return this
-     */
-    public FieldAccessorTable ensureFieldAccessorsInitialized(
-        Class<? extends GeneratedMessageV3> messageClass,
-        Class<? extends Builder> builderClass) {
-      if (initialized) { return this; }
-      synchronized (this) {
-        if (initialized) { return this; }
-        int fieldsSize = fields.length;
-        for (int i = 0; i < fieldsSize; i++) {
-          FieldDescriptor field = descriptor.getFields().get(i);
-          String containingOneofCamelCaseName = null;
-          if (field.getContainingOneof() != null) {
-            containingOneofCamelCaseName =
-                camelCaseNames[fieldsSize + field.getContainingOneof().getIndex()];
-          }
-          if (field.isRepeated()) {
-            if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-              if (field.isMapField()) {
-                fields[i] = new MapFieldAccessor(
-                    field, camelCaseNames[i], messageClass, builderClass);
-              } else {
-                fields[i] = new RepeatedMessageFieldAccessor(
-                    field, camelCaseNames[i], messageClass, builderClass);
-              }
-            } else if (field.getJavaType() == FieldDescriptor.JavaType.ENUM) {
-              fields[i] = new RepeatedEnumFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass);
-            } else {
-              fields[i] = new RepeatedFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass);
-            }
-          } else {
-            if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-              fields[i] = new SingularMessageFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass,
-                  containingOneofCamelCaseName);
-            } else if (field.getJavaType() == FieldDescriptor.JavaType.ENUM) {
-              fields[i] = new SingularEnumFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass,
-                  containingOneofCamelCaseName);
-            } else if (field.getJavaType() == FieldDescriptor.JavaType.STRING) {
-              fields[i] = new SingularStringFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass,
-                  containingOneofCamelCaseName);
-            } else {
-              fields[i] = new SingularFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass,
-                  containingOneofCamelCaseName);
-            }
-          }
-        }
-
-        int oneofsSize = oneofs.length;
-        for (int i = 0; i < oneofsSize; i++) {
-          oneofs[i] = new OneofAccessor(
-              descriptor, camelCaseNames[i + fieldsSize],
-              messageClass, builderClass);
-        }
-        initialized = true;
-        camelCaseNames = null;
-        return this;
-      }
-    }
-
-    private final Descriptor descriptor;
-    private final FieldAccessor[] fields;
-    private String[] camelCaseNames;
-    private final OneofAccessor[] oneofs;
-    private volatile boolean initialized;
-
-    /** Get the FieldAccessor for a particular field. */
-    private FieldAccessor getField(final FieldDescriptor field) {
-      if (field.getContainingType() != descriptor) {
-        throw new IllegalArgumentException(
-          "FieldDescriptor does not match message type.");
-      } else if (field.isExtension()) {
-        // If this type had extensions, it would subclass ExtendableMessage,
-        // which overrides the reflection interface to handle extensions.
-        throw new IllegalArgumentException(
-          "This type does not have extensions.");
-      }
-      return fields[field.getIndex()];
-    }
-
-    /** Get the OneofAccessor for a particular oneof. */
-    private OneofAccessor getOneof(final OneofDescriptor oneof) {
-      if (oneof.getContainingType() != descriptor) {
-        throw new IllegalArgumentException(
-          "OneofDescriptor does not match message type.");
-      }
-      return oneofs[oneof.getIndex()];
-    }
-
-    /**
-     * Abstract interface that provides access to a single field.  This is
-     * implemented differently depending on the field type and cardinality.
-     */
-    private interface FieldAccessor {
-      Object get(GeneratedMessageV3 message);
-      Object get(GeneratedMessageV3.Builder builder);
-      Object getRaw(GeneratedMessageV3 message);
-      Object getRaw(GeneratedMessageV3.Builder builder);
-      void set(Builder builder, Object value);
-      Object getRepeated(GeneratedMessageV3 message, int index);
-      Object getRepeated(GeneratedMessageV3.Builder builder, int index);
-      Object getRepeatedRaw(GeneratedMessageV3 message, int index);
-      Object getRepeatedRaw(GeneratedMessageV3.Builder builder, int index);
-      void setRepeated(Builder builder,
-                       int index, Object value);
-      void addRepeated(Builder builder, Object value);
-      boolean has(GeneratedMessageV3 message);
-      boolean has(GeneratedMessageV3.Builder builder);
-      int getRepeatedCount(GeneratedMessageV3 message);
-      int getRepeatedCount(GeneratedMessageV3.Builder builder);
-      void clear(Builder builder);
-      Message.Builder newBuilder();
-      Message.Builder getBuilder(GeneratedMessageV3.Builder builder);
-      Message.Builder getRepeatedBuilder(GeneratedMessageV3.Builder builder,
-                                         int index);
-    }
-
-    /** OneofAccessor provides access to a single oneof. */
-    private static class OneofAccessor {
-      OneofAccessor(
-          final Descriptor descriptor, final String camelCaseName,
-          final Class<? extends GeneratedMessageV3> messageClass,
-          final Class<? extends Builder> builderClass) {
-        this.descriptor = descriptor;
-        caseMethod =
-            getMethodOrDie(messageClass, "get" + camelCaseName + "Case");
-        caseMethodBuilder =
-            getMethodOrDie(builderClass, "get" + camelCaseName + "Case");
-        clearMethod = getMethodOrDie(builderClass, "clear" + camelCaseName);
-      }
-
-      private final Descriptor descriptor;
-      private final Method caseMethod;
-      private final Method caseMethodBuilder;
-      private final Method clearMethod;
-
-      public boolean has(final GeneratedMessageV3 message) {
-        if (((Internal.EnumLite) invokeOrDie(caseMethod, message)).getNumber() == 0) {
-          return false;
-        }
-        return true;
-      }
-
-      public boolean has(GeneratedMessageV3.Builder builder) {
-        if (((Internal.EnumLite) invokeOrDie(caseMethodBuilder, builder)).getNumber() == 0) {
-          return false;
-        }
-        return true;
-      }
-
-      public FieldDescriptor get(final GeneratedMessageV3 message) {
-        int fieldNumber = ((Internal.EnumLite) invokeOrDie(caseMethod, message)).getNumber();
-        if (fieldNumber > 0) {
-          return descriptor.findFieldByNumber(fieldNumber);
-        }
-        return null;
-      }
-
-      public FieldDescriptor get(GeneratedMessageV3.Builder builder) {
-        int fieldNumber = ((Internal.EnumLite) invokeOrDie(caseMethodBuilder, builder)).getNumber();
-        if (fieldNumber > 0) {
-          return descriptor.findFieldByNumber(fieldNumber);
-        }
-        return null;
-      }
-
-      public void clear(final Builder builder) {
-        invokeOrDie(clearMethod, builder);
-      }
-    }
-
-    private static boolean supportFieldPresence(FileDescriptor file) {
-      return file.getSyntax() == FileDescriptor.Syntax.PROTO2;
-    }
-
-    // ---------------------------------------------------------------
-
-    private static class SingularFieldAccessor implements FieldAccessor {
-      SingularFieldAccessor(
-          final FieldDescriptor descriptor, final String camelCaseName,
-          final Class<? extends GeneratedMessageV3> messageClass,
-          final Class<? extends Builder> builderClass,
-          final String containingOneofCamelCaseName) {
-        field = descriptor;
-        isOneofField = descriptor.getContainingOneof() != null;
-        hasHasMethod = supportFieldPresence(descriptor.getFile())
-            || (!isOneofField && descriptor.getJavaType() == FieldDescriptor.JavaType.MESSAGE);
-        getMethod = getMethodOrDie(messageClass, "get" + camelCaseName);
-        getMethodBuilder = getMethodOrDie(builderClass, "get" + camelCaseName);
-        type = getMethod.getReturnType();
-        setMethod = getMethodOrDie(builderClass, "set" + camelCaseName, type);
-        hasMethod =
-            hasHasMethod ? getMethodOrDie(messageClass, "has" + camelCaseName) : null;
-        hasMethodBuilder =
-            hasHasMethod ? getMethodOrDie(builderClass, "has" + camelCaseName) : null;
-        clearMethod = getMethodOrDie(builderClass, "clear" + camelCaseName);
-        caseMethod = isOneofField ? getMethodOrDie(
-            messageClass, "get" + containingOneofCamelCaseName + "Case") : null;
-        caseMethodBuilder = isOneofField ? getMethodOrDie(
-            builderClass, "get" + containingOneofCamelCaseName + "Case") : null;
-      }
-
-      // Note:  We use Java reflection to call public methods rather than
-      //   access private fields directly as this avoids runtime security
-      //   checks.
-      protected final Class<?> type;
-      protected final Method getMethod;
-      protected final Method getMethodBuilder;
-      protected final Method setMethod;
-      protected final Method hasMethod;
-      protected final Method hasMethodBuilder;
-      protected final Method clearMethod;
-      protected final Method caseMethod;
-      protected final Method caseMethodBuilder;
-      protected final FieldDescriptor field;
-      protected final boolean isOneofField;
-      protected final boolean hasHasMethod;
-
-      private int getOneofFieldNumber(final GeneratedMessageV3 message) {
-        return ((Internal.EnumLite) invokeOrDie(caseMethod, message)).getNumber();
-      }
-
-      private int getOneofFieldNumber(final GeneratedMessageV3.Builder builder) {
-        return ((Internal.EnumLite) invokeOrDie(caseMethodBuilder, builder)).getNumber();
-      }
-
-      @Override
-      public Object get(final GeneratedMessageV3 message) {
-        return invokeOrDie(getMethod, message);
-      }
-      @Override
-      public Object get(GeneratedMessageV3.Builder builder) {
-        return invokeOrDie(getMethodBuilder, builder);
-      }
-      @Override
-      public Object getRaw(final GeneratedMessageV3 message) {
-        return get(message);
-      }
-      @Override
-      public Object getRaw(GeneratedMessageV3.Builder builder) {
-        return get(builder);
-      }
-      @Override
-      public void set(final Builder builder, final Object value) {
-        invokeOrDie(setMethod, builder, value);
-      }
-      @Override
-      public Object getRepeated(final GeneratedMessageV3 message, final int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedField() called on a singular field.");
-      }
-      @Override
-      public Object getRepeatedRaw(final GeneratedMessageV3 message, final int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedFieldRaw() called on a singular field.");
-      }
-      @Override
-      public Object getRepeated(GeneratedMessageV3.Builder builder, int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedField() called on a singular field.");
-      }
-      @Override
-      public Object getRepeatedRaw(GeneratedMessageV3.Builder builder, int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedFieldRaw() called on a singular field.");
-      }
-      @Override
-      public void setRepeated(final Builder builder, final int index, final Object value) {
-        throw new UnsupportedOperationException(
-          "setRepeatedField() called on a singular field.");
-      }
-      @Override
-      public void addRepeated(final Builder builder, final Object value) {
-        throw new UnsupportedOperationException(
-          "addRepeatedField() called on a singular field.");
-      }
-      @Override
-      public boolean has(final GeneratedMessageV3 message) {
-        if (!hasHasMethod) {
-          if (isOneofField) {
-            return getOneofFieldNumber(message) == field.getNumber();
-          }
-          return !get(message).equals(field.getDefaultValue());
-        }
-        return (Boolean) invokeOrDie(hasMethod, message);
-      }
-      @Override
-      public boolean has(GeneratedMessageV3.Builder builder) {
-        if (!hasHasMethod) {
-          if (isOneofField) {
-            return getOneofFieldNumber(builder) == field.getNumber();
-          }
-          return !get(builder).equals(field.getDefaultValue());
-        }
-        return (Boolean) invokeOrDie(hasMethodBuilder, builder);
-      }
-      @Override
-      public int getRepeatedCount(final GeneratedMessageV3 message) {
-        throw new UnsupportedOperationException(
-          "getRepeatedFieldSize() called on a singular field.");
-      }
-      @Override
-      public int getRepeatedCount(GeneratedMessageV3.Builder builder) {
-        throw new UnsupportedOperationException(
-          "getRepeatedFieldSize() called on a singular field.");
-      }
-      @Override
-      public void clear(final Builder builder) {
-        invokeOrDie(clearMethod, builder);
-      }
-      @Override
-      public Message.Builder newBuilder() {
-        throw new UnsupportedOperationException(
-          "newBuilderForField() called on a non-Message type.");
-      }
-      @Override
-      public Message.Builder getBuilder(GeneratedMessageV3.Builder builder) {
-        throw new UnsupportedOperationException(
-          "getFieldBuilder() called on a non-Message type.");
-      }
-      @Override
-      public Message.Builder getRepeatedBuilder(GeneratedMessageV3.Builder builder, int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedFieldBuilder() called on a non-Message type.");
-      }
-    }
-
-    private static class RepeatedFieldAccessor implements FieldAccessor {
-      protected final Class type;
-      protected final Method getMethod;
-      protected final Method getMethodBuilder;
-      protected final Method getRepeatedMethod;
-      protected final Method getRepeatedMethodBuilder;
-      protected final Method setRepeatedMethod;
-      protected final Method addRepeatedMethod;
-      protected final Method getCountMethod;
-      protected final Method getCountMethodBuilder;
-      protected final Method clearMethod;
-
-      RepeatedFieldAccessor(
-          final FieldDescriptor descriptor, final String camelCaseName,
-          final Class<? extends GeneratedMessageV3> messageClass,
-          final Class<? extends Builder> builderClass) {
-        getMethod = getMethodOrDie(messageClass,
-                                   "get" + camelCaseName + "List");
-        getMethodBuilder = getMethodOrDie(builderClass,
-                                   "get" + camelCaseName + "List");
-        getRepeatedMethod =
-            getMethodOrDie(messageClass, "get" + camelCaseName, Integer.TYPE);
-        getRepeatedMethodBuilder =
-            getMethodOrDie(builderClass, "get" + camelCaseName, Integer.TYPE);
-        type = getRepeatedMethod.getReturnType();
-        setRepeatedMethod =
-            getMethodOrDie(builderClass, "set" + camelCaseName,
-                           Integer.TYPE, type);
-        addRepeatedMethod =
-            getMethodOrDie(builderClass, "add" + camelCaseName, type);
-        getCountMethod =
-            getMethodOrDie(messageClass, "get" + camelCaseName + "Count");
-        getCountMethodBuilder =
-            getMethodOrDie(builderClass, "get" + camelCaseName + "Count");
-
-        clearMethod = getMethodOrDie(builderClass, "clear" + camelCaseName);
-      }
-
-      @Override
-      public Object get(final GeneratedMessageV3 message) {
-        return invokeOrDie(getMethod, message);
-      }
-      @Override
-      public Object get(GeneratedMessageV3.Builder builder) {
-        return invokeOrDie(getMethodBuilder, builder);
-      }
-      @Override
-      public Object getRaw(final GeneratedMessageV3 message) {
-        return get(message);
-      }
-      @Override
-      public Object getRaw(GeneratedMessageV3.Builder builder) {
-        return get(builder);
-      }
-      @Override
-      public void set(final Builder builder, final Object value) {
-        // Add all the elements individually.  This serves two purposes:
-        // 1) Verifies that each element has the correct type.
-        // 2) Insures that the caller cannot modify the list later on and
-        //    have the modifications be reflected in the message.
-        clear(builder);
-        for (final Object element : (List<?>) value) {
-          addRepeated(builder, element);
-        }
-      }
-      @Override
-      public Object getRepeated(final GeneratedMessageV3 message, final int index) {
-        return invokeOrDie(getRepeatedMethod, message, index);
-      }
-      @Override
-      public Object getRepeated(GeneratedMessageV3.Builder builder, int index) {
-        return invokeOrDie(getRepeatedMethodBuilder, builder, index);
-      }
-      @Override
-      public Object getRepeatedRaw(GeneratedMessageV3 message, int index) {
-        return getRepeated(message, index);
-      }
-      @Override
-      public Object getRepeatedRaw(GeneratedMessageV3.Builder builder, int index) {
-        return getRepeated(builder, index);
-      }
-      @Override
-      public void setRepeated(final Builder builder, final int index, final Object value) {
-        invokeOrDie(setRepeatedMethod, builder, index, value);
-      }
-      @Override
-      public void addRepeated(final Builder builder, final Object value) {
-        invokeOrDie(addRepeatedMethod, builder, value);
-      }
-      @Override
-      public boolean has(final GeneratedMessageV3 message) {
-        throw new UnsupportedOperationException(
-          "hasField() called on a repeated field.");
-      }
-      @Override
-      public boolean has(GeneratedMessageV3.Builder builder) {
-        throw new UnsupportedOperationException(
-          "hasField() called on a repeated field.");
-      }
-      @Override
-      public int getRepeatedCount(final GeneratedMessageV3 message) {
-        return (Integer) invokeOrDie(getCountMethod, message);
-      }
-      @Override
-      public int getRepeatedCount(GeneratedMessageV3.Builder builder) {
-        return (Integer) invokeOrDie(getCountMethodBuilder, builder);
-      }
-      @Override
-      public void clear(final Builder builder) {
-        invokeOrDie(clearMethod, builder);
-      }
-      @Override
-      public Message.Builder newBuilder() {
-        throw new UnsupportedOperationException(
-          "newBuilderForField() called on a non-Message type.");
-      }
-      @Override
-      public Message.Builder getBuilder(GeneratedMessageV3.Builder builder) {
-        throw new UnsupportedOperationException(
-          "getFieldBuilder() called on a non-Message type.");
-      }
-      @Override
-      public Message.Builder getRepeatedBuilder(GeneratedMessageV3.Builder builder, int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedFieldBuilder() called on a non-Message type.");
-      }
-    }
-
-    private static class MapFieldAccessor implements FieldAccessor {
-      MapFieldAccessor(
-          final FieldDescriptor descriptor, final String camelCaseName,
-          final Class<? extends GeneratedMessageV3> messageClass,
-          final Class<? extends Builder> builderClass) {
-        field = descriptor;
-        Method getDefaultInstanceMethod =
-            getMethodOrDie(messageClass, "getDefaultInstance");
-        MapField defaultMapField = getMapField(
-            (GeneratedMessageV3) invokeOrDie(getDefaultInstanceMethod, null));
-        mapEntryMessageDefaultInstance =
-            defaultMapField.getMapEntryMessageDefaultInstance();
-      }
-
-      private final FieldDescriptor field;
-      private final Message mapEntryMessageDefaultInstance;
-
-      private MapField<?, ?> getMapField(GeneratedMessageV3 message) {
-        return (MapField<?, ?>) message.internalGetMapField(field.getNumber());
-      }
-
-      private MapField<?, ?> getMapField(GeneratedMessageV3.Builder builder) {
-        return (MapField<?, ?>) builder.internalGetMapField(field.getNumber());
-      }
-
-      private MapField<?, ?> getMutableMapField(
-          GeneratedMessageV3.Builder builder) {
-        return (MapField<?, ?>) builder.internalGetMutableMapField(
-            field.getNumber());
-      }
-
-      private Message coerceType(Message value) {
-        if (value == null) {
-          return null;
-        }
-        if (mapEntryMessageDefaultInstance.getClass().isInstance(value)) {
-          return value;
-        }
-        // The value is not the exact right message type.  However, if it
-        // is an alternative implementation of the same type -- e.g. a
-        // DynamicMessage -- we should accept it.  In this case we can make
-        // a copy of the message.
-        return mapEntryMessageDefaultInstance.toBuilder().mergeFrom(value).build();
-      }
-
-      @Override
-      public Object get(GeneratedMessageV3 message) {
-        List result = new ArrayList();
-        for (int i = 0; i < getRepeatedCount(message); i++) {
-          result.add(getRepeated(message, i));
-        }
-        return Collections.unmodifiableList(result);
-      }
-
-      @Override
-      public Object get(Builder builder) {
-        List result = new ArrayList();
-        for (int i = 0; i < getRepeatedCount(builder); i++) {
-          result.add(getRepeated(builder, i));
-        }
-        return Collections.unmodifiableList(result);
-      }
-
-      @Override
-      public Object getRaw(GeneratedMessageV3 message) {
-        return get(message);
-      }
-
-      @Override
-      public Object getRaw(GeneratedMessageV3.Builder builder) {
-        return get(builder);
-      }
-
-      @Override
-      public void set(Builder builder, Object value) {
-        clear(builder);
-        for (Object entry : (List) value) {
-          addRepeated(builder, entry);
-        }
-      }
-
-      @Override
-      public Object getRepeated(GeneratedMessageV3 message, int index) {
-        return getMapField(message).getList().get(index);
-      }
-
-      @Override
-      public Object getRepeated(Builder builder, int index) {
-        return getMapField(builder).getList().get(index);
-      }
-
-      @Override
-      public Object getRepeatedRaw(GeneratedMessageV3 message, int index) {
-        return getRepeated(message, index);
-      }
-
-      @Override
-      public Object getRepeatedRaw(Builder builder, int index) {
-        return getRepeated(builder, index);
-      }
-      
-      @Override
-      public void setRepeated(Builder builder, int index, Object value) {
-        getMutableMapField(builder).getMutableList().set(index, coerceType((Message) value));
-      }
-
-      @Override
-      public void addRepeated(Builder builder, Object value) {
-        getMutableMapField(builder).getMutableList().add(coerceType((Message) value));
-      }
-
-      @Override
-      public boolean has(GeneratedMessageV3 message) {
-        throw new UnsupportedOperationException(
-            "hasField() is not supported for repeated fields.");
-      }
-
-      @Override
-      public boolean has(Builder builder) {
-        throw new UnsupportedOperationException(
-            "hasField() is not supported for repeated fields.");
-      }
-
-      @Override
-      public int getRepeatedCount(GeneratedMessageV3 message) {
-        return getMapField(message).getList().size();
-      }
-
-      @Override
-      public int getRepeatedCount(Builder builder) {
-        return getMapField(builder).getList().size();
-      }
-
-      @Override
-      public void clear(Builder builder) {
-        getMutableMapField(builder).getMutableList().clear();
-      }
-
-      @Override
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Message.Builder newBuilder() {
-        return mapEntryMessageDefaultInstance.newBuilderForType();
-      }
-
-      @Override
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Message.Builder getBuilder(Builder builder) {
-        throw new UnsupportedOperationException(
-            "Nested builder not supported for map fields.");
-      }
-
-      @Override
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Message.Builder getRepeatedBuilder(Builder builder, int index) {
- 

<TRUNCATED>

[26/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessage.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessage.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessage.java
deleted file mode 100644
index 4a0f93a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessage.java
+++ /dev/null
@@ -1,3047 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.ObjectStreamException;
-import java.io.Serializable;
-import java.lang.reflect.InvocationTargetException;
-import java.lang.reflect.Method;
-import java.util.ArrayList;
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-import java.util.TreeMap;
-
-/**
- * All generated protocol message classes extend this class.  This class
- * implements most of the Message and Builder interfaces using Java reflection.
- * Users can ignore this class and pretend that generated messages implement
- * the Message interface directly.
- *
- * @author kenton@google.com Kenton Varda
- */
-public abstract class GeneratedMessage extends AbstractMessage
-    implements Serializable {
-  private static final long serialVersionUID = 1L;
-
-  /**
-   * For testing. Allows a test to disable the optimization that avoids using
-   * field builders for nested messages until they are requested. By disabling
-   * this optimization, existing tests can be reused to test the field builders.
-   */
-  protected static boolean alwaysUseFieldBuilders = false;
-
-  /** For use by generated code only.  */
-  protected UnknownFieldSet unknownFields;
-
-  protected GeneratedMessage() {
-    unknownFields = UnknownFieldSet.getDefaultInstance();
-  }
-
-  protected GeneratedMessage(Builder<?> builder) {
-    unknownFields = builder.getUnknownFields();
-  }
-
-  @Override
-  public Parser<? extends GeneratedMessage> getParserForType() {
-    throw new UnsupportedOperationException(
-        "This is supposed to be overridden by subclasses.");
-  }
-
- /**
-  * For testing. Allows a test to disable the optimization that avoids using
-  * field builders for nested messages until they are requested. By disabling
-  * this optimization, existing tests can be reused to test the field builders.
-  * See {@link RepeatedFieldBuilder} and {@link SingleFieldBuilder}.
-  */
-  static void enableAlwaysUseFieldBuildersForTesting() {
-    alwaysUseFieldBuilders = true;
-  }
-
-  /**
-   * Get the FieldAccessorTable for this type.  We can't have the message
-   * class pass this in to the constructor because of bootstrapping trouble
-   * with DescriptorProtos.
-   */
-  protected abstract FieldAccessorTable internalGetFieldAccessorTable();
-
-  @Override
-  public Descriptor getDescriptorForType() {
-    return internalGetFieldAccessorTable().descriptor;
-  }
-
-  /**
-   * Internal helper to return a modifiable map containing all the fields.
-   * The returned Map is modifialbe so that the caller can add additional
-   * extension fields to implement {@link #getAllFields()}.
-   *
-   * @param getBytesForString whether to generate ByteString for string fields
-   */
-  private Map<FieldDescriptor, Object> getAllFieldsMutable(
-      boolean getBytesForString) {
-    final TreeMap<FieldDescriptor, Object> result =
-      new TreeMap<FieldDescriptor, Object>();
-    final Descriptor descriptor = internalGetFieldAccessorTable().descriptor;
-    final List<FieldDescriptor> fields = descriptor.getFields();
-
-    for (int i = 0; i < fields.size(); i++) {
-      FieldDescriptor field = fields.get(i);
-      final OneofDescriptor oneofDescriptor = field.getContainingOneof();
-
-      /*
-       * If the field is part of a Oneof, then at maximum one field in the Oneof is set
-       * and it is not repeated. There is no need to iterate through the others.
-       */
-      if (oneofDescriptor != null) {
-        // Skip other fields in the Oneof we know are not set
-        i += oneofDescriptor.getFieldCount() - 1;
-        if (!hasOneof(oneofDescriptor)) {
-          // If no field is set in the Oneof, skip all the fields in the Oneof
-          continue;
-        }
-        // Get the pointer to the only field which is set in the Oneof
-        field = getOneofFieldDescriptor(oneofDescriptor);
-      } else {
-        // If we are not in a Oneof, we need to check if the field is set and if it is repeated
-        if (field.isRepeated()) {
-          final List<?> value = (List<?>) getField(field);
-          if (!value.isEmpty()) {
-            result.put(field, value);
-          }
-          continue;
-        }
-        if (!hasField(field)) {
-          continue;
-        }
-      }
-      // Add the field to the map
-      if (getBytesForString && field.getJavaType() == FieldDescriptor.JavaType.STRING) {
-        result.put(field, getFieldRaw(field));
-      } else {
-        result.put(field, getField(field));
-      }
-    }
-    return result;
-  }
-
-  @Override
-  public boolean isInitialized() {
-    for (final FieldDescriptor field : getDescriptorForType().getFields()) {
-      // Check that all required fields are present.
-      if (field.isRequired()) {
-        if (!hasField(field)) {
-          return false;
-        }
-      }
-      // Check that embedded messages are initialized.
-      if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-        if (field.isRepeated()) {
-          @SuppressWarnings("unchecked") final
-          List<Message> messageList = (List<Message>) getField(field);
-          for (final Message element : messageList) {
-            if (!element.isInitialized()) {
-              return false;
-            }
-          }
-        } else {
-          if (hasField(field) && !((Message) getField(field)).isInitialized()) {
-            return false;
-          }
-        }
-      }
-    }
-
-    return true;
-  }
-
-  @Override
-  public Map<FieldDescriptor, Object> getAllFields() {
-    return Collections.unmodifiableMap(
-        getAllFieldsMutable(/* getBytesForString = */ false));
-  }
-
-  /**
-   * Returns a collection of all the fields in this message which are set
-   * and their corresponding values.  A singular ("required" or "optional")
-   * field is set iff hasField() returns true for that field.  A "repeated"
-   * field is set iff getRepeatedFieldCount() is greater than zero.  The
-   * values are exactly what would be returned by calling
-   * {@link #getFieldRaw(Descriptors.FieldDescriptor)} for each field.  The map
-   * is guaranteed to be a sorted map, so iterating over it will return fields
-   * in order by field number.
-   */
-  Map<FieldDescriptor, Object> getAllFieldsRaw() {
-    return Collections.unmodifiableMap(
-        getAllFieldsMutable(/* getBytesForString = */ true));
-  }
-
-  @Override
-  public boolean hasOneof(final OneofDescriptor oneof) {
-    return internalGetFieldAccessorTable().getOneof(oneof).has(this);
-  }
-
-  @Override
-  public FieldDescriptor getOneofFieldDescriptor(final OneofDescriptor oneof) {
-    return internalGetFieldAccessorTable().getOneof(oneof).get(this);
-  }
-
-  @Override
-  public boolean hasField(final FieldDescriptor field) {
-    return internalGetFieldAccessorTable().getField(field).has(this);
-  }
-
-  @Override
-  public Object getField(final FieldDescriptor field) {
-    return internalGetFieldAccessorTable().getField(field).get(this);
-  }
-
-  /**
-   * Obtains the value of the given field, or the default value if it is
-   * not set.  For primitive fields, the boxed primitive value is returned.
-   * For enum fields, the EnumValueDescriptor for the value is returned. For
-   * embedded message fields, the sub-message is returned.  For repeated
-   * fields, a java.util.List is returned. For present string fields, a
-   * ByteString is returned representing the bytes that the field contains.
-   */
-  Object getFieldRaw(final FieldDescriptor field) {
-    return internalGetFieldAccessorTable().getField(field).getRaw(this);
-  }
-
-  @Override
-  public int getRepeatedFieldCount(final FieldDescriptor field) {
-    return internalGetFieldAccessorTable().getField(field)
-      .getRepeatedCount(this);
-  }
-
-  @Override
-  public Object getRepeatedField(final FieldDescriptor field, final int index) {
-    return internalGetFieldAccessorTable().getField(field)
-      .getRepeated(this, index);
-  }
-
-  @Override
-  public UnknownFieldSet getUnknownFields() {
-    throw new UnsupportedOperationException(
-        "This is supposed to be overridden by subclasses.");
-  }
-
-  /**
-   * Called by subclasses to parse an unknown field.
-   * @return {@code true} unless the tag is an end-group tag.
-   */
-  protected boolean parseUnknownField(
-      CodedInputStream input,
-      UnknownFieldSet.Builder unknownFields,
-      ExtensionRegistryLite extensionRegistry,
-      int tag) throws IOException {
-    return unknownFields.mergeFieldFrom(tag, input);
-  }
-
-  protected static <M extends Message> M parseWithIOException(Parser<M> parser, InputStream input)
-      throws IOException {
-    try {
-      return parser.parseFrom(input);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseWithIOException(Parser<M> parser, InputStream input,
-      ExtensionRegistryLite extensions) throws IOException {
-    try {
-      return parser.parseFrom(input, extensions);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseWithIOException(Parser<M> parser,
-      CodedInputStream input) throws IOException {
-    try {
-      return parser.parseFrom(input);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseWithIOException(Parser<M> parser,
-      CodedInputStream input, ExtensionRegistryLite extensions) throws IOException {
-    try {
-      return parser.parseFrom(input, extensions);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseDelimitedWithIOException(Parser<M> parser,
-      InputStream input) throws IOException {
-    try {
-      return parser.parseDelimitedFrom(input);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  protected static <M extends Message> M parseDelimitedWithIOException(Parser<M> parser,
-      InputStream input, ExtensionRegistryLite extensions) throws IOException {
-    try {
-      return parser.parseDelimitedFrom(input, extensions);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.unwrapIOException();
-    }
-  }
-
-  @Override
-  public void writeTo(final CodedOutputStream output) throws IOException {
-    MessageReflection.writeMessageTo(this, getAllFieldsRaw(), output, false);
-  }
-
-  @Override
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) {
-      return size;
-    }
-
-    memoizedSize = MessageReflection.getSerializedSize(
-        this, getAllFieldsRaw());
-    return memoizedSize;
-  }
-
-
-
-  /**
-   * Used by parsing constructors in generated classes.
-   */
-  protected void makeExtensionsImmutable() {
-    // Noop for messages without extensions.
-  }
-
-  /**
-   * TODO(xiaofeng): remove this after b/29368482 is fixed. We need to move this
-   * interface to AbstractMessage in order to versioning GeneratedMessage but
-   * this move breaks binary compatibility for AppEngine. After AppEngine is
-   * fixed we can exlude this from google3.
-   */
-  protected interface BuilderParent extends AbstractMessage.BuilderParent {}
-
-  /**
-   * TODO(xiaofeng): remove this together with GeneratedMessage.BuilderParent.
-   */
-  protected abstract Message.Builder newBuilderForType(BuilderParent parent);
-
-  @Override
-  protected Message.Builder newBuilderForType(final AbstractMessage.BuilderParent parent) {
-    return newBuilderForType(new BuilderParent() {
-      @Override
-      public void markDirty() {
-        parent.markDirty();
-      }
-    });
-  }
-
-
-  @SuppressWarnings("unchecked")
-  public abstract static class Builder <BuilderType extends Builder<BuilderType>>
-      extends AbstractMessage.Builder<BuilderType> {
-
-    private BuilderParent builderParent;
-
-    private BuilderParentImpl meAsParent;
-
-    // Indicates that we've built a message and so we are now obligated
-    // to dispatch dirty invalidations. See GeneratedMessage.BuilderListener.
-    private boolean isClean;
-
-    private UnknownFieldSet unknownFields =
-        UnknownFieldSet.getDefaultInstance();
-
-    protected Builder() {
-      this(null);
-    }
-
-    protected Builder(BuilderParent builderParent) {
-      this.builderParent = builderParent;
-    }
-
-    @Override
-    void dispose() {
-      builderParent = null;
-    }
-
-    /**
-     * Called by the subclass when a message is built.
-     */
-    protected void onBuilt() {
-      if (builderParent != null) {
-        markClean();
-      }
-    }
-
-    /**
-     * Called by the subclass or a builder to notify us that a message was
-     * built and may be cached and therefore invalidations are needed.
-     */
-    @Override
-    protected void markClean() {
-      this.isClean = true;
-    }
-
-    /**
-     * Gets whether invalidations are needed
-     *
-     * @return whether invalidations are needed
-     */
-    protected boolean isClean() {
-      return isClean;
-    }
-
-    @Override
-    public BuilderType clone() {
-      BuilderType builder =
-          (BuilderType) getDefaultInstanceForType().newBuilderForType();
-      builder.mergeFrom(buildPartial());
-      return builder;
-    }
-
-    /**
-     * Called by the initialization and clear code paths to allow subclasses to
-     * reset any of their builtin fields back to the initial values.
-     */
-    @Override
-    public BuilderType clear() {
-      unknownFields = UnknownFieldSet.getDefaultInstance();
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /**
-     * Get the FieldAccessorTable for this type.  We can't have the message
-     * class pass this in to the constructor because of bootstrapping trouble
-     * with DescriptorProtos.
-     */
-    protected abstract FieldAccessorTable internalGetFieldAccessorTable();
-
-    @Override
-    public Descriptor getDescriptorForType() {
-      return internalGetFieldAccessorTable().descriptor;
-    }
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFields() {
-      return Collections.unmodifiableMap(getAllFieldsMutable());
-    }
-
-    /** Internal helper which returns a mutable map. */
-    private Map<FieldDescriptor, Object> getAllFieldsMutable() {
-      final TreeMap<FieldDescriptor, Object> result =
-        new TreeMap<FieldDescriptor, Object>();
-      final Descriptor descriptor = internalGetFieldAccessorTable().descriptor;
-      final List<FieldDescriptor> fields = descriptor.getFields();
-
-      for (int i = 0; i < fields.size(); i++) {
-        FieldDescriptor field = fields.get(i);
-        final OneofDescriptor oneofDescriptor = field.getContainingOneof();
-
-        /*
-         * If the field is part of a Oneof, then at maximum one field in the Oneof is set
-         * and it is not repeated. There is no need to iterate through the others.
-         */
-        if (oneofDescriptor != null) {
-          // Skip other fields in the Oneof we know are not set
-          i += oneofDescriptor.getFieldCount() - 1;
-          if (!hasOneof(oneofDescriptor)) {
-            // If no field is set in the Oneof, skip all the fields in the Oneof
-            continue;
-          }
-          // Get the pointer to the only field which is set in the Oneof
-          field = getOneofFieldDescriptor(oneofDescriptor);
-        } else {
-          // If we are not in a Oneof, we need to check if the field is set and if it is repeated
-          if (field.isRepeated()) {
-            final List<?> value = (List<?>) getField(field);
-            if (!value.isEmpty()) {
-              result.put(field, value);
-            }
-            continue;
-          }
-          if (!hasField(field)) {
-            continue;
-          }
-        }
-        // Add the field to the map
-        result.put(field, getField(field));
-      }
-      return result;
-    }
-
-    @Override
-    public Message.Builder newBuilderForField(final FieldDescriptor field) {
-      return internalGetFieldAccessorTable().getField(field).newBuilder();
-    }
-
-    @Override
-    public Message.Builder getFieldBuilder(final FieldDescriptor field) {
-      return internalGetFieldAccessorTable().getField(field).getBuilder(this);
-    }
-
-    @Override
-    public Message.Builder getRepeatedFieldBuilder(final FieldDescriptor field, int index) {
-      return internalGetFieldAccessorTable().getField(field).getRepeatedBuilder(
-          this, index);
-    }
-
-    @Override
-    public boolean hasOneof(final OneofDescriptor oneof) {
-      return internalGetFieldAccessorTable().getOneof(oneof).has(this);
-    }
-
-    @Override
-    public FieldDescriptor getOneofFieldDescriptor(final OneofDescriptor oneof) {
-      return internalGetFieldAccessorTable().getOneof(oneof).get(this);
-    }
-
-    @Override
-    public boolean hasField(final FieldDescriptor field) {
-      return internalGetFieldAccessorTable().getField(field).has(this);
-    }
-
-    @Override
-    public Object getField(final FieldDescriptor field) {
-      Object object = internalGetFieldAccessorTable().getField(field).get(this);
-      if (field.isRepeated()) {
-        // The underlying list object is still modifiable at this point.
-        // Make sure not to expose the modifiable list to the caller.
-        return Collections.unmodifiableList((List) object);
-      } else {
-        return object;
-      }
-    }
-
-    @Override
-    public BuilderType setField(final FieldDescriptor field, final Object value) {
-      internalGetFieldAccessorTable().getField(field).set(this, value);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType clearField(final FieldDescriptor field) {
-      internalGetFieldAccessorTable().getField(field).clear(this);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType clearOneof(final OneofDescriptor oneof) {
-      internalGetFieldAccessorTable().getOneof(oneof).clear(this);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public int getRepeatedFieldCount(final FieldDescriptor field) {
-      return internalGetFieldAccessorTable().getField(field)
-          .getRepeatedCount(this);
-    }
-
-    @Override
-    public Object getRepeatedField(final FieldDescriptor field, final int index) {
-      return internalGetFieldAccessorTable().getField(field)
-          .getRepeated(this, index);
-    }
-
-    @Override
-    public BuilderType setRepeatedField(
-        final FieldDescriptor field, final int index, final Object value) {
-      internalGetFieldAccessorTable().getField(field)
-        .setRepeated(this, index, value);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType addRepeatedField(final FieldDescriptor field, final Object value) {
-      internalGetFieldAccessorTable().getField(field).addRepeated(this, value);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType setUnknownFields(final UnknownFieldSet unknownFields) {
-      this.unknownFields = unknownFields;
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType mergeUnknownFields(
-        final UnknownFieldSet unknownFields) {
-      this.unknownFields =
-        UnknownFieldSet.newBuilder(this.unknownFields)
-                       .mergeFrom(unknownFields)
-                       .build();
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    @Override
-    public boolean isInitialized() {
-      for (final FieldDescriptor field : getDescriptorForType().getFields()) {
-        // Check that all required fields are present.
-        if (field.isRequired()) {
-          if (!hasField(field)) {
-            return false;
-          }
-        }
-        // Check that embedded messages are initialized.
-        if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-          if (field.isRepeated()) {
-            @SuppressWarnings("unchecked") final
-            List<Message> messageList = (List<Message>) getField(field);
-            for (final Message element : messageList) {
-              if (!element.isInitialized()) {
-                return false;
-              }
-            }
-          } else {
-            if (hasField(field) &&
-                !((Message) getField(field)).isInitialized()) {
-              return false;
-            }
-          }
-        }
-      }
-      return true;
-    }
-
-    @Override
-    public final UnknownFieldSet getUnknownFields() {
-      return unknownFields;
-    }
-
-    /**
-     * Called by subclasses to parse an unknown field.
-     * @return {@code true} unless the tag is an end-group tag.
-     */
-    protected boolean parseUnknownField(
-        final CodedInputStream input,
-        final UnknownFieldSet.Builder unknownFields,
-        final ExtensionRegistryLite extensionRegistry,
-        final int tag) throws IOException {
-      return unknownFields.mergeFieldFrom(tag, input);
-    }
-
-    /**
-     * Implementation of {@link BuilderParent} for giving to our children. This
-     * small inner class makes it so we don't publicly expose the BuilderParent
-     * methods.
-     */
-    private class BuilderParentImpl implements BuilderParent {
-
-      @Override
-      public void markDirty() {
-        onChanged();
-      }
-    }
-
-    /**
-     * Gets the {@link BuilderParent} for giving to our children.
-     * @return The builder parent for our children.
-     */
-    protected BuilderParent getParentForChildren() {
-      if (meAsParent == null) {
-        meAsParent = new BuilderParentImpl();
-      }
-      return meAsParent;
-    }
-
-    /**
-     * Called when a the builder or one of its nested children has changed
-     * and any parent should be notified of its invalidation.
-     */
-    protected final void onChanged() {
-      if (isClean && builderParent != null) {
-        builderParent.markDirty();
-
-        // Don't keep dispatching invalidations until build is called again.
-        isClean = false;
-      }
-    }
-
-    /**
-     * Gets the map field with the given field number. This method should be
-     * overridden in the generated message class if the message contains map
-     * fields.
-     *
-     * Unlike other field types, reflection support for map fields can't be
-     * implemented based on generated public API because we need to access a
-     * map field as a list in reflection API but the generated API only allows
-     * us to access it as a map. This method returns the underlying map field
-     * directly and thus enables us to access the map field as a list.
-     */
-    @SuppressWarnings({"unused", "rawtypes"})
-    protected MapField internalGetMapField(int fieldNumber) {
-      // Note that we can't use descriptor names here because this method will
-      // be called when descriptor is being initialized.
-      throw new RuntimeException(
-          "No map fields found in " + getClass().getName());
-    }
-
-    /** Like {@link #internalGetMapField} but return a mutable version. */
-    @SuppressWarnings({"unused", "rawtypes"})
-    protected MapField internalGetMutableMapField(int fieldNumber) {
-      // Note that we can't use descriptor names here because this method will
-      // be called when descriptor is being initialized.
-      throw new RuntimeException(
-          "No map fields found in " + getClass().getName());
-    }
-  }
-
-  // =================================================================
-  // Extensions-related stuff
-
-  public interface ExtendableMessageOrBuilder<
-      MessageType extends ExtendableMessage> extends MessageOrBuilder {
-    // Re-define for return type covariance.
-    @Override
-    Message getDefaultInstanceForType();
-
-    /** Check if a singular extension is present. */
-    <Type> boolean hasExtension(
-        ExtensionLite<MessageType, Type> extension);
-
-    /** Get the number of elements in a repeated extension. */
-    <Type> int getExtensionCount(
-        ExtensionLite<MessageType, List<Type>> extension);
-
-    /** Get the value of an extension. */
-    <Type> Type getExtension(
-        ExtensionLite<MessageType, Type> extension);
-
-    /** Get one element of a repeated extension. */
-    <Type> Type getExtension(
-        ExtensionLite<MessageType, List<Type>> extension,
-        int index);
-
-    /** Check if a singular extension is present. */
-    <Type> boolean hasExtension(
-        Extension<MessageType, Type> extension);
-    /** Check if a singular extension is present. */
-    <Type> boolean hasExtension(
-        GeneratedExtension<MessageType, Type> extension);
-    /** Get the number of elements in a repeated extension. */
-    <Type> int getExtensionCount(
-        Extension<MessageType, List<Type>> extension);
-    /** Get the number of elements in a repeated extension. */
-    <Type> int getExtensionCount(
-        GeneratedExtension<MessageType, List<Type>> extension);
-    /** Get the value of an extension. */
-    <Type> Type getExtension(
-        Extension<MessageType, Type> extension);
-    /** Get the value of an extension. */
-    <Type> Type getExtension(
-        GeneratedExtension<MessageType, Type> extension);
-    /** Get one element of a repeated extension. */
-    <Type> Type getExtension(
-        Extension<MessageType, List<Type>> extension,
-        int index);
-    /** Get one element of a repeated extension. */
-    <Type> Type getExtension(
-        GeneratedExtension<MessageType, List<Type>> extension,
-        int index);
-  }
-
-  /**
-   * Generated message classes for message types that contain extension ranges
-   * subclass this.
-   *
-   * <p>This class implements type-safe accessors for extensions.  They
-   * implement all the same operations that you can do with normal fields --
-   * e.g. "has", "get", and "getCount" -- but for extensions.  The extensions
-   * are identified using instances of the class {@link GeneratedExtension};
-   * the protocol compiler generates a static instance of this class for every
-   * extension in its input.  Through the magic of generics, all is made
-   * type-safe.
-   *
-   * <p>For example, imagine you have the {@code .proto} file:
-   *
-   * <pre>
-   * option java_class = "MyProto";
-   *
-   * message Foo {
-   *   extensions 1000 to max;
-   * }
-   *
-   * extend Foo {
-   *   optional int32 bar;
-   * }
-   * </pre>
-   *
-   * <p>Then you might write code like:
-   *
-   * <pre>
-   * MyProto.Foo foo = getFoo();
-   * int i = foo.getExtension(MyProto.bar);
-   * </pre>
-   *
-   * <p>See also {@link ExtendableBuilder}.
-   */
-  public abstract static class ExtendableMessage<
-        MessageType extends ExtendableMessage>
-      extends GeneratedMessage
-      implements ExtendableMessageOrBuilder<MessageType> {
-
-    private static final long serialVersionUID = 1L;
-
-    private final FieldSet<FieldDescriptor> extensions;
-
-    protected ExtendableMessage() {
-      this.extensions = FieldSet.newFieldSet();
-    }
-
-    protected ExtendableMessage(
-        ExtendableBuilder<MessageType, ?> builder) {
-      super(builder);
-      this.extensions = builder.buildExtensions();
-    }
-
-    private void verifyExtensionContainingType(
-        final Extension<MessageType, ?> extension) {
-      if (extension.getDescriptor().getContainingType() !=
-          getDescriptorForType()) {
-        // This can only happen if someone uses unchecked operations.
-        throw new IllegalArgumentException(
-          "Extension is for type \"" +
-          extension.getDescriptor().getContainingType().getFullName() +
-          "\" which does not match message type \"" +
-          getDescriptorForType().getFullName() + "\".");
-      }
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final ExtensionLite<MessageType, Type> extensionLite) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      return extensions.hasField(extension.getDescriptor());
-    }
-
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final ExtensionLite<MessageType, List<Type>> extensionLite) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      return extensions.getRepeatedFieldCount(descriptor);
-    }
-
-    /** Get the value of an extension. */
-    @Override
-    @SuppressWarnings("unchecked")
-    public final <Type> Type getExtension(final ExtensionLite<MessageType, Type> extensionLite) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      FieldDescriptor descriptor = extension.getDescriptor();
-      final Object value = extensions.getField(descriptor);
-      if (value == null) {
-        if (descriptor.isRepeated()) {
-          return (Type) Collections.emptyList();
-        } else if (descriptor.getJavaType() ==
-                   FieldDescriptor.JavaType.MESSAGE) {
-          return (Type) extension.getMessageDefaultInstance();
-        } else {
-          return (Type) extension.fromReflectionType(
-              descriptor.getDefaultValue());
-        }
-      } else {
-        return (Type) extension.fromReflectionType(value);
-      }
-    }
-
-    /** Get one element of a repeated extension. */
-    @Override
-    @SuppressWarnings("unchecked")
-    public final <Type> Type getExtension(
-        final ExtensionLite<MessageType, List<Type>> extensionLite, final int index) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      FieldDescriptor descriptor = extension.getDescriptor();
-      return (Type) extension.singularFromReflectionType(
-          extensions.getRepeatedField(descriptor, index));
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final Extension<MessageType, Type> extension) {
-      return hasExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(
-        final GeneratedExtension<MessageType, Type> extension) {
-      return hasExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final Extension<MessageType, List<Type>> extension) {
-      return getExtensionCount((ExtensionLite<MessageType, List<Type>>) extension);
-    }
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final GeneratedExtension<MessageType, List<Type>> extension) {
-      return getExtensionCount((ExtensionLite<MessageType, List<Type>>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(final Extension<MessageType, Type> extension) {
-      return getExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final GeneratedExtension<MessageType, Type> extension) {
-      return getExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get one element of a repeated extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final Extension<MessageType, List<Type>> extension, final int index) {
-      return getExtension((ExtensionLite<MessageType, List<Type>>) extension, index);
-    }
-    /** Get one element of a repeated extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final GeneratedExtension<MessageType, List<Type>> extension, final int index) {
-      return getExtension((ExtensionLite<MessageType, List<Type>>) extension, index);
-    }
-
-    /** Called by subclasses to check if all extensions are initialized. */
-    protected boolean extensionsAreInitialized() {
-      return extensions.isInitialized();
-    }
-
-    @Override
-    public boolean isInitialized() {
-      return super.isInitialized() && extensionsAreInitialized();
-    }
-
-    @Override
-    protected boolean parseUnknownField(
-        CodedInputStream input,
-        UnknownFieldSet.Builder unknownFields,
-        ExtensionRegistryLite extensionRegistry,
-        int tag) throws IOException {
-      return MessageReflection.mergeFieldFrom(
-          input, unknownFields, extensionRegistry, getDescriptorForType(),
-          new MessageReflection.ExtensionAdapter(extensions), tag);
-    }
-
-
-    /**
-     * Used by parsing constructors in generated classes.
-     */
-    @Override
-    protected void makeExtensionsImmutable() {
-      extensions.makeImmutable();
-    }
-
-    /**
-     * Used by subclasses to serialize extensions.  Extension ranges may be
-     * interleaved with field numbers, but we must write them in canonical
-     * (sorted by field number) order.  ExtensionWriter helps us write
-     * individual ranges of extensions at once.
-     */
-    protected class ExtensionWriter {
-      // Imagine how much simpler this code would be if Java iterators had
-      // a way to get the next element without advancing the iterator.
-
-      private final Iterator<Map.Entry<FieldDescriptor, Object>> iter =
-        extensions.iterator();
-      private Map.Entry<FieldDescriptor, Object> next;
-      private final boolean messageSetWireFormat;
-
-      private ExtensionWriter(final boolean messageSetWireFormat) {
-        if (iter.hasNext()) {
-          next = iter.next();
-        }
-        this.messageSetWireFormat = messageSetWireFormat;
-      }
-
-      public void writeUntil(final int end, final CodedOutputStream output)
-                             throws IOException {
-        while (next != null && next.getKey().getNumber() < end) {
-          FieldDescriptor descriptor = next.getKey();
-          if (messageSetWireFormat && descriptor.getLiteJavaType() ==
-                  WireFormat.JavaType.MESSAGE &&
-              !descriptor.isRepeated()) {
-            if (next instanceof LazyField.LazyEntry<?>) {
-              output.writeRawMessageSetExtension(descriptor.getNumber(),
-                  ((LazyField.LazyEntry<?>) next).getField().toByteString());
-            } else {
-              output.writeMessageSetExtension(descriptor.getNumber(),
-                                              (Message) next.getValue());
-            }
-          } else {
-            // TODO(xiangl): Taken care of following code, it may cause
-            // problem when we use LazyField for normal fields/extensions.
-            // Due to the optional field can be duplicated at the end of
-            // serialized bytes, which will make the serialized size change
-            // after lazy field parsed. So when we use LazyField globally,
-            // we need to change the following write method to write cached
-            // bytes directly rather than write the parsed message.
-            FieldSet.writeField(descriptor, next.getValue(), output);
-          }
-          if (iter.hasNext()) {
-            next = iter.next();
-          } else {
-            next = null;
-          }
-        }
-      }
-    }
-
-    protected ExtensionWriter newExtensionWriter() {
-      return new ExtensionWriter(false);
-    }
-    protected ExtensionWriter newMessageSetExtensionWriter() {
-      return new ExtensionWriter(true);
-    }
-
-    /** Called by subclasses to compute the size of extensions. */
-    protected int extensionsSerializedSize() {
-      return extensions.getSerializedSize();
-    }
-    protected int extensionsSerializedSizeAsMessageSet() {
-      return extensions.getMessageSetSerializedSize();
-    }
-
-    // ---------------------------------------------------------------
-    // Reflection
-
-    protected Map<FieldDescriptor, Object> getExtensionFields() {
-      return extensions.getAllFields();
-    }
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFields() {
-      final Map<FieldDescriptor, Object> result =
-          super.getAllFieldsMutable(/* getBytesForString = */ false);
-      result.putAll(getExtensionFields());
-      return Collections.unmodifiableMap(result);
-    }
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFieldsRaw() {
-      final Map<FieldDescriptor, Object> result =
-          super.getAllFieldsMutable(/* getBytesForString = */ false);
-      result.putAll(getExtensionFields());
-      return Collections.unmodifiableMap(result);
-    }
-
-    @Override
-    public boolean hasField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.hasField(field);
-      } else {
-        return super.hasField(field);
-      }
-    }
-
-    @Override
-    public Object getField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        final Object value = extensions.getField(field);
-        if (value == null) {
-          if (field.isRepeated()) {
-            return Collections.emptyList();
-          } else if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-            // Lacking an ExtensionRegistry, we have no way to determine the
-            // extension's real type, so we return a DynamicMessage.
-            return DynamicMessage.getDefaultInstance(field.getMessageType());
-          } else {
-            return field.getDefaultValue();
-          }
-        } else {
-          return value;
-        }
-      } else {
-        return super.getField(field);
-      }
-    }
-
-    @Override
-    public int getRepeatedFieldCount(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.getRepeatedFieldCount(field);
-      } else {
-        return super.getRepeatedFieldCount(field);
-      }
-    }
-
-    @Override
-    public Object getRepeatedField(final FieldDescriptor field,
-                                   final int index) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.getRepeatedField(field, index);
-      } else {
-        return super.getRepeatedField(field, index);
-      }
-    }
-
-    private void verifyContainingType(final FieldDescriptor field) {
-      if (field.getContainingType() != getDescriptorForType()) {
-        throw new IllegalArgumentException(
-          "FieldDescriptor does not match message type.");
-      }
-    }
-  }
-
-  /**
-   * Generated message builders for message types that contain extension ranges
-   * subclass this.
-   *
-   * <p>This class implements type-safe accessors for extensions.  They
-   * implement all the same operations that you can do with normal fields --
-   * e.g. "get", "set", and "add" -- but for extensions.  The extensions are
-   * identified using instances of the class {@link GeneratedExtension}; the
-   * protocol compiler generates a static instance of this class for every
-   * extension in its input.  Through the magic of generics, all is made
-   * type-safe.
-   *
-   * <p>For example, imagine you have the {@code .proto} file:
-   *
-   * <pre>
-   * option java_class = "MyProto";
-   *
-   * message Foo {
-   *   extensions 1000 to max;
-   * }
-   *
-   * extend Foo {
-   *   optional int32 bar;
-   * }
-   * </pre>
-   *
-   * <p>Then you might write code like:
-   *
-   * <pre>
-   * MyProto.Foo foo =
-   *   MyProto.Foo.newBuilder()
-   *     .setExtension(MyProto.bar, 123)
-   *     .build();
-   * </pre>
-   *
-   * <p>See also {@link ExtendableMessage}.
-   */
-  @SuppressWarnings("unchecked")
-  public abstract static class ExtendableBuilder<
-        MessageType extends ExtendableMessage,
-        BuilderType extends ExtendableBuilder<MessageType, BuilderType>>
-      extends Builder<BuilderType>
-      implements ExtendableMessageOrBuilder<MessageType> {
-
-    private FieldSet<FieldDescriptor> extensions = FieldSet.emptySet();
-
-    protected ExtendableBuilder() {}
-
-    protected ExtendableBuilder(
-        BuilderParent parent) {
-      super(parent);
-    }
-
-    // For immutable message conversion.
-    void internalSetExtensionSet(FieldSet<FieldDescriptor> extensions) {
-      this.extensions = extensions;
-    }
-
-    @Override
-    public BuilderType clear() {
-      extensions = FieldSet.emptySet();
-      return super.clear();
-    }
-
-    // This is implemented here only to work around an apparent bug in the
-    // Java compiler and/or build system.  See bug #1898463.  The mere presence
-    // of this clone() implementation makes it go away.
-    @Override
-    public BuilderType clone() {
-      return super.clone();
-    }
-
-    private void ensureExtensionsIsMutable() {
-      if (extensions.isImmutable()) {
-        extensions = extensions.clone();
-      }
-    }
-
-    private void verifyExtensionContainingType(
-        final Extension<MessageType, ?> extension) {
-      if (extension.getDescriptor().getContainingType() !=
-          getDescriptorForType()) {
-        // This can only happen if someone uses unchecked operations.
-        throw new IllegalArgumentException(
-          "Extension is for type \"" +
-          extension.getDescriptor().getContainingType().getFullName() +
-          "\" which does not match message type \"" +
-          getDescriptorForType().getFullName() + "\".");
-      }
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final ExtensionLite<MessageType, Type> extensionLite) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      return extensions.hasField(extension.getDescriptor());
-    }
-
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final ExtensionLite<MessageType, List<Type>> extensionLite) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      return extensions.getRepeatedFieldCount(descriptor);
-    }
-
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(final ExtensionLite<MessageType, Type> extensionLite) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      FieldDescriptor descriptor = extension.getDescriptor();
-      final Object value = extensions.getField(descriptor);
-      if (value == null) {
-        if (descriptor.isRepeated()) {
-          return (Type) Collections.emptyList();
-        } else if (descriptor.getJavaType() ==
-                   FieldDescriptor.JavaType.MESSAGE) {
-          return (Type) extension.getMessageDefaultInstance();
-        } else {
-          return (Type) extension.fromReflectionType(
-              descriptor.getDefaultValue());
-        }
-      } else {
-        return (Type) extension.fromReflectionType(value);
-      }
-    }
-
-    /** Get one element of a repeated extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final ExtensionLite<MessageType, List<Type>> extensionLite, final int index) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      FieldDescriptor descriptor = extension.getDescriptor();
-      return (Type) extension.singularFromReflectionType(
-          extensions.getRepeatedField(descriptor, index));
-    }
-
-    /** Set the value of an extension. */
-    public final <Type> BuilderType setExtension(
-        final ExtensionLite<MessageType, Type> extensionLite,
-        final Type value) {
-      Extension<MessageType, Type> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      ensureExtensionsIsMutable();
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      extensions.setField(descriptor, extension.toReflectionType(value));
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /** Set the value of one element of a repeated extension. */
-    public final <Type> BuilderType setExtension(
-        final ExtensionLite<MessageType, List<Type>> extensionLite,
-        final int index, final Type value) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      ensureExtensionsIsMutable();
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      extensions.setRepeatedField(
-        descriptor, index,
-        extension.singularToReflectionType(value));
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /** Append a value to a repeated extension. */
-    public final <Type> BuilderType addExtension(
-        final ExtensionLite<MessageType, List<Type>> extensionLite,
-        final Type value) {
-      Extension<MessageType, List<Type>> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      ensureExtensionsIsMutable();
-      final FieldDescriptor descriptor = extension.getDescriptor();
-      extensions.addRepeatedField(
-          descriptor, extension.singularToReflectionType(value));
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /** Clear an extension. */
-    public final <Type> BuilderType clearExtension(
-        final ExtensionLite<MessageType, ?> extensionLite) {
-      Extension<MessageType, ?> extension = checkNotLite(extensionLite);
-
-      verifyExtensionContainingType(extension);
-      ensureExtensionsIsMutable();
-      extensions.clearField(extension.getDescriptor());
-      onChanged();
-      return (BuilderType) this;
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final Extension<MessageType, Type> extension) {
-      return hasExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(
-        final GeneratedExtension<MessageType, Type> extension) {
-      return hasExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final Extension<MessageType, List<Type>> extension) {
-      return getExtensionCount((ExtensionLite<MessageType, List<Type>>) extension);
-    }
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final GeneratedExtension<MessageType, List<Type>> extension) {
-      return getExtensionCount((ExtensionLite<MessageType, List<Type>>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(final Extension<MessageType, Type> extension) {
-      return getExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final GeneratedExtension<MessageType, Type> extension) {
-      return getExtension((ExtensionLite<MessageType, Type>) extension);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final Extension<MessageType, List<Type>> extension, final int index) {
-      return getExtension((ExtensionLite<MessageType, List<Type>>) extension, index);
-    }
-    /** Get the value of an extension. */
-    @Override
-    public final <Type> Type getExtension(
-        final GeneratedExtension<MessageType, List<Type>> extension, final int index) {
-      return getExtension((ExtensionLite<MessageType, List<Type>>) extension, index);
-    }
-    /** Set the value of an extension. */
-    public final <Type> BuilderType setExtension(
-        final Extension<MessageType, Type> extension, final Type value) {
-      return setExtension((ExtensionLite<MessageType, Type>) extension, value);
-    }
-    /** Set the value of an extension. */
-    public <Type> BuilderType setExtension(
-        final GeneratedExtension<MessageType, Type> extension, final Type value) {
-      return setExtension((ExtensionLite<MessageType, Type>) extension, value);
-    }
-    /** Set the value of one element of a repeated extension. */
-    public final <Type> BuilderType setExtension(
-        final Extension<MessageType, List<Type>> extension,
-        final int index, final Type value) {
-      return setExtension((ExtensionLite<MessageType, List<Type>>) extension, index, value);
-    }
-    /** Set the value of one element of a repeated extension. */
-    public <Type> BuilderType setExtension(
-        final GeneratedExtension<MessageType, List<Type>> extension,
-        final int index, final Type value) {
-      return setExtension((ExtensionLite<MessageType, List<Type>>) extension, index, value);
-    }
-    /** Append a value to a repeated extension. */
-    public final <Type> BuilderType addExtension(
-        final Extension<MessageType, List<Type>> extension, final Type value) {
-      return addExtension((ExtensionLite<MessageType, List<Type>>) extension, value);
-    }
-    /** Append a value to a repeated extension. */
-    public <Type> BuilderType addExtension(
-        final GeneratedExtension<MessageType, List<Type>> extension, final Type value) {
-      return addExtension((ExtensionLite<MessageType, List<Type>>) extension, value);
-    }
-    /** Clear an extension. */
-    public final <Type> BuilderType clearExtension(
-        final Extension<MessageType, ?> extension) {
-      return clearExtension((ExtensionLite<MessageType, ?>) extension);
-    }
-    /** Clear an extension. */
-    public <Type> BuilderType clearExtension(
-        final GeneratedExtension<MessageType, ?> extension) {
-      return clearExtension((ExtensionLite<MessageType, ?>) extension);
-    }
-
-    /** Called by subclasses to check if all extensions are initialized. */
-    protected boolean extensionsAreInitialized() {
-      return extensions.isInitialized();
-    }
-
-    /**
-     * Called by the build code path to create a copy of the extensions for
-     * building the message.
-     */
-    private FieldSet<FieldDescriptor> buildExtensions() {
-      extensions.makeImmutable();
-      return extensions;
-    }
-
-    @Override
-    public boolean isInitialized() {
-      return super.isInitialized() && extensionsAreInitialized();
-    }
-
-    /**
-     * Called by subclasses to parse an unknown field or an extension.
-     * @return {@code true} unless the tag is an end-group tag.
-     */
-    @Override
-    protected boolean parseUnknownField(
-        final CodedInputStream input,
-        final UnknownFieldSet.Builder unknownFields,
-        final ExtensionRegistryLite extensionRegistry,
-        final int tag) throws IOException {
-      return MessageReflection.mergeFieldFrom(
-          input, unknownFields, extensionRegistry, getDescriptorForType(),
-          new MessageReflection.BuilderAdapter(this), tag);
-    }
-
-    // ---------------------------------------------------------------
-    // Reflection
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFields() {
-      final Map<FieldDescriptor, Object> result = super.getAllFieldsMutable();
-      result.putAll(extensions.getAllFields());
-      return Collections.unmodifiableMap(result);
-    }
-
-    @Override
-    public Object getField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        final Object value = extensions.getField(field);
-        if (value == null) {
-          if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-            // Lacking an ExtensionRegistry, we have no way to determine the
-            // extension's real type, so we return a DynamicMessage.
-            return DynamicMessage.getDefaultInstance(field.getMessageType());
-          } else {
-            return field.getDefaultValue();
-          }
-        } else {
-          return value;
-        }
-      } else {
-        return super.getField(field);
-      }
-    }
-
-    @Override
-    public int getRepeatedFieldCount(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.getRepeatedFieldCount(field);
-      } else {
-        return super.getRepeatedFieldCount(field);
-      }
-    }
-
-    @Override
-    public Object getRepeatedField(final FieldDescriptor field,
-                                   final int index) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.getRepeatedField(field, index);
-      } else {
-        return super.getRepeatedField(field, index);
-      }
-    }
-
-    @Override
-    public boolean hasField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        return extensions.hasField(field);
-      } else {
-        return super.hasField(field);
-      }
-    }
-
-    @Override
-    public BuilderType setField(final FieldDescriptor field,
-                                final Object value) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        ensureExtensionsIsMutable();
-        extensions.setField(field, value);
-        onChanged();
-        return (BuilderType) this;
-      } else {
-        return super.setField(field, value);
-      }
-    }
-
-    @Override
-    public BuilderType clearField(final FieldDescriptor field) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        ensureExtensionsIsMutable();
-        extensions.clearField(field);
-        onChanged();
-        return (BuilderType) this;
-      } else {
-        return super.clearField(field);
-      }
-    }
-
-    @Override
-    public BuilderType setRepeatedField(final FieldDescriptor field,
-                                        final int index, final Object value) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        ensureExtensionsIsMutable();
-        extensions.setRepeatedField(field, index, value);
-        onChanged();
-        return (BuilderType) this;
-      } else {
-        return super.setRepeatedField(field, index, value);
-      }
-    }
-
-    @Override
-    public BuilderType addRepeatedField(final FieldDescriptor field,
-                                        final Object value) {
-      if (field.isExtension()) {
-        verifyContainingType(field);
-        ensureExtensionsIsMutable();
-        extensions.addRepeatedField(field, value);
-        onChanged();
-        return (BuilderType) this;
-      } else {
-        return super.addRepeatedField(field, value);
-      }
-    }
-
-    protected final void mergeExtensionFields(final ExtendableMessage other) {
-      ensureExtensionsIsMutable();
-      extensions.mergeFrom(other.extensions);
-      onChanged();
-    }
-
-    private void verifyContainingType(final FieldDescriptor field) {
-      if (field.getContainingType() != getDescriptorForType()) {
-        throw new IllegalArgumentException(
-          "FieldDescriptor does not match message type.");
-      }
-    }
-  }
-
-  // -----------------------------------------------------------------
-
-  /**
-   * Gets the descriptor for an extension. The implementation depends on whether
-   * the extension is scoped in the top level of a file or scoped in a Message.
-   */
-  static interface ExtensionDescriptorRetriever {
-    FieldDescriptor getDescriptor();
-  }
-
-  /** For use by generated code only. */
-  public static <ContainingType extends Message, Type>
-      GeneratedExtension<ContainingType, Type>
-      newMessageScopedGeneratedExtension(final Message scope,
-                                         final int descriptorIndex,
-                                         final Class singularType,
-                                         final Message defaultInstance) {
-    // For extensions scoped within a Message, we use the Message to resolve
-    // the outer class's descriptor, from which the extension descriptor is
-    // obtained.
-    return new GeneratedExtension<ContainingType, Type>(
-        new CachedDescriptorRetriever() {
-          @Override
-          public FieldDescriptor loadDescriptor() {
-            return scope.getDescriptorForType().getExtensions().get(descriptorIndex);
-          }
-        },
-        singularType,
-        defaultInstance,
-        Extension.ExtensionType.IMMUTABLE);
-  }
-
-  /** For use by generated code only. */
-  public static <ContainingType extends Message, Type>
-     GeneratedExtension<ContainingType, Type>
-     newFileScopedGeneratedExtension(final Class singularType,
-                                     final Message defaultInstance) {
-    // For extensions scoped within a file, we rely on the outer class's
-    // static initializer to call internalInit() on the extension when the
-    // descriptor is available.
-    return new GeneratedExtension<ContainingType, Type>(
-        null,  // ExtensionDescriptorRetriever is initialized in internalInit();
-        singularType,
-        defaultInstance,
-        Extension.ExtensionType.IMMUTABLE);
-  }
-
-  private abstract static class CachedDescriptorRetriever
-      implements ExtensionDescriptorRetriever {
-    private volatile FieldDescriptor descriptor;
-    protected abstract FieldDescriptor loadDescriptor();
-
-    @Override
-    public FieldDescriptor getDescriptor() {
-      if (descriptor == null) {
-        synchronized (this) {
-          if (descriptor == null) {
-            descriptor = loadDescriptor();
-          }
-        }
-      }
-      return descriptor;
-    }
-  }
-
-  /**
-   * Used in proto1 generated code only.
-   *
-   * After enabling bridge, we can define proto2 extensions (the extended type
-   * is a proto2 mutable message) in a proto1 .proto file. For these extensions
-   * we should generate proto2 GeneratedExtensions.
-   */
-  public static <ContainingType extends Message, Type>
-      GeneratedExtension<ContainingType, Type>
-      newMessageScopedGeneratedExtension(
-          final Message scope, final String name,
-          final Class singularType, final Message defaultInstance) {
-    // For extensions scoped within a Message, we use the Message to resolve
-    // the outer class's descriptor, from which the extension descriptor is
-    // obtained.
-    return new GeneratedExtension<ContainingType, Type>(
-        new CachedDescriptorRetriever() {
-          @Override
-          protected FieldDescriptor loadDescriptor() {
-            return scope.getDescriptorForType().findFieldByName(name);
-          }
-        },
-        singularType,
-        defaultInstance,
-        Extension.ExtensionType.MUTABLE);
-  }
-
-  /**
-   * Used in proto1 generated code only.
-   *
-   * After enabling bridge, we can define proto2 extensions (the extended type
-   * is a proto2 mutable message) in a proto1 .proto file. For these extensions
-   * we should generate proto2 GeneratedExtensions.
-   */
-  public static <ContainingType extends Message, Type>
-     GeneratedExtension<ContainingType, Type>
-     newFileScopedGeneratedExtension(
-         final Class singularType, final Message defaultInstance,
-         final String descriptorOuterClass, final String extensionName) {
-    // For extensions scoped within a file, we load the descriptor outer
-    // class and rely on it to get the FileDescriptor which then can be
-    // used to obtain the extension's FieldDescriptor.
-    return new GeneratedExtension<ContainingType, Type>(
-        new CachedDescriptorRetriever() {
-          @Override
-          protected FieldDescriptor loadDescriptor() {
-            try {
-              Class clazz = singularType.getClassLoader().loadClass(descriptorOuterClass);
-              FileDescriptor file = (FileDescriptor) clazz.getField("descriptor").get(null);
-              return file.findExtensionByName(extensionName);
-            } catch (Exception e) {
-              throw new RuntimeException(
-                  "Cannot load descriptors: "
-                      + descriptorOuterClass
-                      + " is not a valid descriptor class name",
-                  e);
-            }
-          }
-        },
-        singularType,
-        defaultInstance,
-        Extension.ExtensionType.MUTABLE);
-  }
-
-  /**
-   * Type used to represent generated extensions.  The protocol compiler
-   * generates a static singleton instance of this class for each extension.
-   *
-   * <p>For example, imagine you have the {@code .proto} file:
-   *
-   * <pre>
-   * option java_class = "MyProto";
-   *
-   * message Foo {
-   *   extensions 1000 to max;
-   * }
-   *
-   * extend Foo {
-   *   optional int32 bar;
-   * }
-   * </pre>
-   *
-   * <p>Then, {@code MyProto.Foo.bar} has type
-   * {@code GeneratedExtension<MyProto.Foo, Integer>}.
-   *
-   * <p>In general, users should ignore the details of this type, and simply use
-   * these static singletons as parameters to the extension accessors defined
-   * in {@link ExtendableMessage} and {@link ExtendableBuilder}.
-   */
-  public static class GeneratedExtension<
-      ContainingType extends Message, Type> extends
-      Extension<ContainingType, Type> {
-    // TODO(kenton):  Find ways to avoid using Java reflection within this
-    //   class.  Also try to avoid suppressing unchecked warnings.
-
-    // We can't always initialize the descriptor of a GeneratedExtension when
-    // we first construct it due to initialization order difficulties (namely,
-    // the descriptor may not have been constructed yet, since it is often
-    // constructed by the initializer of a separate module).
-    //
-    // In the case of nested extensions, we initialize the
-    // ExtensionDescriptorRetriever with an instance that uses the scoping
-    // Message's default instance to retrieve the extension's descriptor.
-    //
-    // In the case of non-nested extensions, we initialize the
-    // ExtensionDescriptorRetriever to null and rely on the outer class's static
-    // initializer to call internalInit() after the descriptor has been parsed.
-    GeneratedExtension(ExtensionDescriptorRetriever descriptorRetriever,
-        Class singularType,
-        Message messageDefaultInstance,
-        ExtensionType extensionType) {
-      if (Message.class.isAssignableFrom(singularType) &&
-          !singularType.isInstance(messageDefaultInstance)) {
-        throw new IllegalArgumentException(
-            "Bad messageDefaultInstance for " + singularType.getName());
-      }
-      this.descriptorRetriever = descriptorRetriever;
-      this.singularType = singularType;
-      this.messageDefaultInstance = messageDefaultInstance;
-
-      if (ProtocolMessageEnum.class.isAssignableFrom(singularType)) {
-        this.enumValueOf = getMethodOrDie(singularType, "valueOf",
-                                          EnumValueDescriptor.class);
-        this.enumGetValueDescriptor =
-            getMethodOrDie(singularType, "getValueDescriptor");
-      } else {
-        this.enumValueOf = null;
-        this.enumGetValueDescriptor = null;
-      }
-      this.extensionType = extensionType;
-    }
-
-    /** For use by generated code only. */
-    public void internalInit(final FieldDescriptor descriptor) {
-      if (descriptorRetriever != null) {
-        throw new IllegalStateException("Already initialized.");
-      }
-      descriptorRetriever =
-          new ExtensionDescriptorRetriever() {
-            @Override
-            public FieldDescriptor getDescriptor() {
-              return descriptor;
-            }
-          };
-    }
-
-    private ExtensionDescriptorRetriever descriptorRetriever;
-    private final Class singularType;
-    private final Message messageDefaultInstance;
-    private final Method enumValueOf;
-    private final Method enumGetValueDescriptor;
-    private final ExtensionType extensionType;
-
-    @Override
-    public FieldDescriptor getDescriptor() {
-      if (descriptorRetriever == null) {
-        throw new IllegalStateException(
-            "getDescriptor() called before internalInit()");
-      }
-      return descriptorRetriever.getDescriptor();
-    }
-
-    /**
-     * If the extension is an embedded message or group, returns the default
-     * instance of the message.
-     */
-    @Override
-    public Message getMessageDefaultInstance() {
-      return messageDefaultInstance;
-    }
-
-    @Override
-    protected ExtensionType getExtensionType() {
-      return extensionType;
-    }
-
-    /**
-     * Convert from the type used by the reflection accessors to the type used
-     * by native accessors.  E.g., for enums, the reflection accessors use
-     * EnumValueDescriptors but the native accessors use the generated enum
-     * type.
-     */
-    @Override
-    @SuppressWarnings("unchecked")
-    protected Object fromReflectionType(final Object value) {
-      FieldDescriptor descriptor = getDescriptor();
-      if (descriptor.isRepeated()) {
-        if (descriptor.getJavaType() == FieldDescriptor.JavaType.MESSAGE ||
-            descriptor.getJavaType() == FieldDescriptor.JavaType.ENUM) {
-          // Must convert the whole list.
-          final List result = new ArrayList();
-          for (final Object element : (List) value) {
-            result.add(singularFromReflectionType(element));
-          }
-          return result;
-        } else {
-          return value;
-        }
-      } else {
-        return singularFromReflectionType(value);
-      }
-    }
-
-    /**
-     * Like {@link #fromReflectionType(Object)}, but if the type is a repeated
-     * type, this converts a single element.
-     */
-    @Override
-    protected Object singularFromReflectionType(final Object value) {
-      FieldDescriptor descriptor = getDescriptor();
-      switch (descriptor.getJavaType()) {
-        case MESSAGE:
-          if (singularType.isInstance(value)) {
-            return value;
-          } else {
-            return messageDefaultInstance.newBuilderForType()
-                .mergeFrom((Message) value).build();
-          }
-        case ENUM:
-          return invokeOrDie(enumValueOf, null, (EnumValueDescriptor) value);
-        default:
-          return value;
-      }
-    }
-
-    /**
-     * Convert from the type used by the native accessors to the type used
-     * by reflection accessors.  E.g., for enums, the reflection accessors use
-     * EnumValueDescriptors but the native accessors use the generated enum
-     * type.
-     */
-    @Override
-    @SuppressWarnings("unchecked")
-    protected Object toReflectionType(final Object value) {
-      FieldDescriptor descriptor = getDescriptor();
-      if (descriptor.isRepeated()) {
-        if (descriptor.getJavaType() == FieldDescriptor.JavaType.ENUM) {
-          // Must convert the whole list.
-          final List result = new ArrayList();
-          for (final Object element : (List) value) {
-            result.add(singularToReflectionType(element));
-          }
-          return result;
-        } else {
-          return value;
-        }
-      } else {
-        return singularToReflectionType(value);
-      }
-    }
-
-    /**
-     * Like {@link #toReflectionType(Object)}, but if the type is a repeated
-     * type, this converts a single element.
-     */
-    @Override
-    protected Object singularToReflectionType(final Object value) {
-      FieldDescriptor descriptor = getDescriptor();
-      switch (descriptor.getJavaType()) {
-        case ENUM:
-          return invokeOrDie(enumGetValueDescriptor, value);
-        default:
-          return value;
-      }
-    }
-
-    @Override
-    public int getNumber() {
-      return getDescriptor().getNumber();
-    }
-
-    @Override
-    public WireFormat.FieldType getLiteType() {
-      return getDescriptor().getLiteType();
-    }
-
-    @Override
-    public boolean isRepeated() {
-      return getDescriptor().isRepeated();
-    }
-
-    @Override
-    @SuppressWarnings("unchecked")
-    public Type getDefaultValue() {
-      if (isRepeated()) {
-        return (Type) Collections.emptyList();
-      }
-      if (getDescriptor().getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-        return (Type) messageDefaultInstance;
-      }
-      return (Type) singularFromReflectionType(
-          getDescriptor().getDefaultValue());
-    }
-  }
-
-  // =================================================================
-
-  /** Calls Class.getMethod and throws a RuntimeException if it fails. */
-  @SuppressWarnings("unchecked")
-  private static Method getMethodOrDie(
-      final Class clazz, final String name, final Class... params) {
-    try {
-      return clazz.getMethod(name, params);
-    } catch (NoSuchMethodException e) {
-      throw new RuntimeException(
-        "Generated message class \"" + clazz.getName() +
-        "\" missing method \"" + name + "\".", e);
-    }
-  }
-
-  /** Calls invoke and throws a RuntimeException if it fails. */
-  private static Object invokeOrDie(
-      final Method method, final Object object, final Object... params) {
-    try {
-      return method.invoke(object, params);
-    } catch (IllegalAccessException e) {
-      throw new RuntimeException(
-        "Couldn't use Java reflection to implement protocol message " +
-        "reflection.", e);
-    } catch (InvocationTargetException e) {
-      final Throwable cause = e.getCause();
-      if (cause instanceof RuntimeException) {
-        throw (RuntimeException) cause;
-      } else if (cause instanceof Error) {
-        throw (Error) cause;
-      } else {
-        throw new RuntimeException(
-          "Unexpected exception thrown by generated accessor method.", cause);
-      }
-    }
-  }
-
-  /**
-   * Gets the map field with the given field number. This method should be
-   * overridden in the generated message class if the message contains map
-   * fields.
-   *
-   * Unlike other field types, reflection support for map fields can't be
-   * implemented based on generated public API because we need to access a
-   * map field as a list in reflection API but the generated API only allows
-   * us to access it as a map. This method returns the underlying map field
-   * directly and thus enables us to access the map field as a list.
-   */
-  @SuppressWarnings({"rawtypes", "unused"})
-  protected MapField internalGetMapField(int fieldNumber) {
-    // Note that we can't use descriptor names here because this method will
-    // be called when descriptor is being initialized.
-    throw new RuntimeException(
-        "No map fields found in " + getClass().getName());
-  }
-
-  /**
-   * Users should ignore this class.  This class provides the implementation
-   * with access to the fields of a message object using Java reflection.
-   */
-  public static final class FieldAccessorTable {
-
-    /**
-     * Construct a FieldAccessorTable for a particular message class.  Only
-     * one FieldAccessorTable should ever be constructed per class.
-     *
-     * @param descriptor     The type's descriptor.
-     * @param camelCaseNames The camelcase names of all fields in the message.
-     *                       These are used to derive the accessor method names.
-     * @param messageClass   The message type.
-     * @param builderClass   The builder type.
-     */
-    public FieldAccessorTable(
-        final Descriptor descriptor,
-        final String[] camelCaseNames,
-        final Class<? extends GeneratedMessage> messageClass,
-        final Class<? extends Builder> builderClass) {
-      this(descriptor, camelCaseNames);
-      ensureFieldAccessorsInitialized(messageClass, builderClass);
-    }
-
-    /**
-     * Construct a FieldAccessorTable for a particular message class without
-     * initializing FieldAccessors.
-     */
-    public FieldAccessorTable(
-        final Descriptor descriptor,
-        final String[] camelCaseNames) {
-      this.descriptor = descriptor;
-      this.camelCaseNames = camelCaseNames;
-      fields = new FieldAccessor[descriptor.getFields().size()];
-      oneofs = new OneofAccessor[descriptor.getOneofs().size()];
-      initialized = false;
-    }
-
-    private boolean isMapFieldEnabled(FieldDescriptor field) {
-      boolean result = true;
-      return result;
-    }
-
-    /**
-     * Ensures the field accessors are initialized. This method is thread-safe.
-     *
-     * @param messageClass   The message type.
-     * @param builderClass   The builder type.
-     * @return this
-     */
-    public FieldAccessorTable ensureFieldAccessorsInitialized(
-        Class<? extends GeneratedMessage> messageClass,
-        Class<? extends Builder> builderClass) {
-      if (initialized) { return this; }
-      synchronized (this) {
-        if (initialized) { return this; }
-        int fieldsSize = fields.length;
-        for (int i = 0; i < fieldsSize; i++) {
-          FieldDescriptor field = descriptor.getFields().get(i);
-          String containingOneofCamelCaseName = null;
-          if (field.getContainingOneof() != null) {
-            containingOneofCamelCaseName =
-                camelCaseNames[fieldsSize + field.getContainingOneof().getIndex()];
-          }
-          if (field.isRepeated()) {
-            if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-              if (field.isMapField() && isMapFieldEnabled(field)) {
-                fields[i] = new MapFieldAccessor(
-                    field, camelCaseNames[i], messageClass, builderClass);
-              } else {
-                fields[i] = new RepeatedMessageFieldAccessor(
-                    field, camelCaseNames[i], messageClass, builderClass);
-              }
-            } else if (field.getJavaType() == FieldDescriptor.JavaType.ENUM) {
-              fields[i] = new RepeatedEnumFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass);
-            } else {
-              fields[i] = new RepeatedFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass);
-            }
-          } else {
-            if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-              fields[i] = new SingularMessageFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass,
-                  containingOneofCamelCaseName);
-            } else if (field.getJavaType() == FieldDescriptor.JavaType.ENUM) {
-              fields[i] = new SingularEnumFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass,
-                  containingOneofCamelCaseName);
-            } else if (field.getJavaType() == FieldDescriptor.JavaType.STRING) {
-              fields[i] = new SingularStringFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass,
-                  containingOneofCamelCaseName);
-            } else {
-              fields[i] = new SingularFieldAccessor(
-                  field, camelCaseNames[i], messageClass, builderClass,
-                  containingOneofCamelCaseName);
-            }
-          }
-        }
-
-        int oneofsSize = oneofs.length;
-        for (int i = 0; i < oneofsSize; i++) {
-          oneofs[i] = new OneofAccessor(
-              descriptor, camelCaseNames[i + fieldsSize],
-              messageClass, builderClass);
-        }
-        initialized = true;
-        camelCaseNames = null;
-        return this;
-      }
-    }
-
-    private final Descriptor descriptor;
-    private final FieldAccessor[] fields;
-    private String[] camelCaseNames;
-    private final OneofAccessor[] oneofs;
-    private volatile boolean initialized;
-
-    /** Get the FieldAccessor for a particular field. */
-    private FieldAccessor getField(final FieldDescriptor field) {
-      if (field.getContainingType() != descriptor) {
-        throw new IllegalArgumentException(
-          "FieldDescriptor does not match message type.");
-      } else if (field.isExtension()) {
-        // If this type had extensions, it would subclass ExtendableMessage,
-        // which overrides the reflection interface to handle extensions.
-        throw new IllegalArgumentException(
-          "This type does not have extensions.");
-      }
-      return fields[field.getIndex()];
-    }
-
-    /** Get the OneofAccessor for a particular oneof. */
-    private OneofAccessor getOneof(final OneofDescriptor oneof) {
-      if (oneof.getContainingType() != descriptor) {
-        throw new IllegalArgumentException(
-          "OneofDescriptor does not match message type.");
-      }
-      return oneofs[oneof.getIndex()];
-    }
-
-    /**
-     * Abstract interface that provides access to a single field.  This is
-     * implemented differently depending on the field type and cardinality.
-     */
-    private interface FieldAccessor {
-      Object get(GeneratedMessage message);
-      Object get(GeneratedMessage.Builder builder);
-      Object getRaw(GeneratedMessage message);
-      Object getRaw(GeneratedMessage.Builder builder);
-      void set(Builder builder, Object value);
-      Object getRepeated(GeneratedMessage message, int index);
-      Object getRepeated(GeneratedMessage.Builder builder, int index);
-      Object getRepeatedRaw(GeneratedMessage message, int index);
-      Object getRepeatedRaw(GeneratedMessage.Builder builder, int index);
-      void setRepeated(Builder builder,
-                       int index, Object value);
-      void addRepeated(Builder builder, Object value);
-      boolean has(GeneratedMessage message);
-      boolean has(GeneratedMessage.Builder builder);
-      int getRepeatedCount(GeneratedMessage message);
-      int getRepeatedCount(GeneratedMessage.Builder builder);
-      void clear(Builder builder);
-      Message.Builder newBuilder();
-      Message.Builder getBuilder(GeneratedMessage.Builder builder);
-      Message.Builder getRepeatedBuilder(GeneratedMessage.Builder builder,
-                                         int index);
-    }
-
-    /** OneofAccessor provides access to a single oneof. */
-    private static class OneofAccessor {
-      OneofAccessor(
-          final Descriptor descriptor, final String camelCaseName,
-          final Class<? extends GeneratedMessage> messageClass,
-          final Class<? extends Builder> builderClass) {
-        this.descriptor = descriptor;
-        caseMethod =
-            getMethodOrDie(messageClass, "get" + camelCaseName + "Case");
-        caseMethodBuilder =
-            getMethodOrDie(builderClass, "get" + camelCaseName + "Case");
-        clearMethod = getMethodOrDie(builderClass, "clear" + camelCaseName);
-      }
-
-      private final Descriptor descriptor;
-      private final Method caseMethod;
-      private final Method caseMethodBuilder;
-      private final Method clearMethod;
-
-      public boolean has(final GeneratedMessage message) {
-        if (((Internal.EnumLite) invokeOrDie(caseMethod, message)).getNumber() == 0) {
-          return false;
-        }
-        return true;
-      }
-
-      public boolean has(GeneratedMessage.Builder builder) {
-        if (((Internal.EnumLite) invokeOrDie(caseMethodBuilder, builder)).getNumber() == 0) {
-          return false;
-        }
-        return true;
-      }
-
-      public FieldDescriptor get(final GeneratedMessage message) {
-        int fieldNumber = ((Internal.EnumLite) invokeOrDie(caseMethod, message)).getNumber();
-        if (fieldNumber > 0) {
-          return descriptor.findFieldByNumber(fieldNumber);
-        }
-        return null;
-      }
-
-      public FieldDescriptor get(GeneratedMessage.Builder builder) {
-        int fieldNumber = ((Internal.EnumLite) invokeOrDie(caseMethodBuilder, builder)).getNumber();
-        if (fieldNumber > 0) {
-          return descriptor.findFieldByNumber(fieldNumber);
-        }
-        return null;
-      }
-
-      public void clear(final Builder builder) {
-        invokeOrDie(clearMethod, builder);
-      }
-    }
-
-    private static boolean supportFieldPresence(FileDescriptor file) {
-      return file.getSyntax() == FileDescriptor.Syntax.PROTO2;
-    }
-
-    // ---------------------------------------------------------------
-
-    private static class SingularFieldAccessor implements FieldAccessor {
-      SingularFieldAccessor(
-          final FieldDescriptor descriptor, final String camelCaseName,
-          final Class<? extends GeneratedMessage> messageClass,
-          final Class<? extends Builder> builderClass,
-          final String containingOneofCamelCaseName) {
-        field = descriptor;
-        isOneofField = descriptor.getContainingOneof() != null;
-        hasHasMethod = supportFieldPresence(descriptor.getFile())
-            || (!isOneofField && descriptor.getJavaType() == FieldDescriptor.JavaType.MESSAGE);
-        getMethod = getMethodOrDie(messageClass, "get" + camelCaseName);
-        getMethodBuilder = getMethodOrDie(builderClass, "get" + camelCaseName);
-        type = getMethod.getReturnType();
-        setMethod = getMethodOrDie(builderClass, "set" + camelCaseName, type);
-        hasMethod =
-            hasHasMethod ? getMethodOrDie(messageClass, "has" + camelCaseName) : null;
-        hasMethodBuilder =
-            hasHasMethod ? getMethodOrDie(builderClass, "has" + camelCaseName) : null;
-        clearMethod = getMethodOrDie(builderClass, "clear" + camelCaseName);
-        caseMethod = isOneofField ? getMethodOrDie(
-            messageClass, "get" + containingOneofCamelCaseName + "Case") : null;
-        caseMethodBuilder = isOneofField ? getMethodOrDie(
-            builderClass, "get" + containingOneofCamelCaseName + "Case") : null;
-      }
-
-      // Note:  We use Java reflection to call public methods rather than
-      //   access private fields directly as this avoids runtime security
-      //   checks.
-      protected final Class<?> type;
-      protected final Method getMethod;
-      protected final Method getMethodBuilder;
-      protected final Method setMethod;
-      protected final Method hasMethod;
-      protected final Method hasMethodBuilder;
-      protected final Method clearMethod;
-      protected final Method caseMethod;
-      protected final Method caseMethodBuilder;
-      protected final FieldDescriptor field;
-      protected final boolean isOneofField;
-      protected final boolean hasHasMethod;
-
-      private int getOneofFieldNumber(final GeneratedMessage message) {
-        return ((Internal.EnumLite) invokeOrDie(caseMethod, message)).getNumber();
-      }
-
-      private int getOneofFieldNumber(final GeneratedMessage.Builder builder) {
-        return ((Internal.EnumLite) invokeOrDie(caseMethodBuilder, builder)).getNumber();
-      }
-
-      @Override
-      public Object get(final GeneratedMessage message) {
-        return invokeOrDie(getMethod, message);
-      }
-      @Override
-      public Object get(GeneratedMessage.Builder builder) {
-        return invokeOrDie(getMethodBuilder, builder);
-      }
-      @Override
-      public Object getRaw(final GeneratedMessage message) {
-        return get(message);
-      }
-      @Override
-      public Object getRaw(GeneratedMessage.Builder builder) {
-        return get(builder);
-      }
-      @Override
-      public void set(final Builder builder, final Object value) {
-        invokeOrDie(setMethod, builder, value);
-      }
-      @Override
-      public Object getRepeated(final GeneratedMessage message, final int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedField() called on a singular field.");
-      }
-      @Override
-      public Object getRepeatedRaw(final GeneratedMessage message, final int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedFieldRaw() called on a singular field.");
-      }
-      @Override
-      public Object getRepeated(GeneratedMessage.Builder builder, int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedField() called on a singular field.");
-      }
-      @Override
-      public Object getRepeatedRaw(GeneratedMessage.Builder builder, int index) {
-        throw new UnsupportedOperationException(
-          "getRepeatedFieldRaw() called on a singular field.");
-      }
-      @Override
-      public void setRepeated(final Builder builder, final int index, final Object value) {
-        throw new UnsupportedOperationException(
-          "setRepeatedField() called on a singular field.");
-      }
-      @Override
-      public void addRepeated(final Builder builder, final Object value) {
-        throw new UnsupportedOperationException(
-          "addRepeatedField() called on a singular field.");
-      }
-      @Override
-      public boolean has(final GeneratedMessage message) {
-        if (!hasHasMethod) {
-          if (isOneofField) {
-            return getOneofFieldNumber(message) == field.getNumber();
-          }
-          return !get(message).equals(field.getDefaultValue());
-        }
-        return (Boolean) invokeOrDie(hasMethod, message);
-      }

<TRUNCATED>

[43/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java
----------------------------------------------------------------------
diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java
deleted file mode 100644
index 86c88c6..0000000
--- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/SecureBulkLoadProtos.java
+++ /dev/null
@@ -1,2088 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: SecureBulkLoad.proto
-
-package org.apache.hadoop.hbase.protobuf.generated;
-
-public final class SecureBulkLoadProtos {
-  private SecureBulkLoadProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface SecureBulkLoadHFilesRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> 
-        getFamilyPathList();
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index);
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    int getFamilyPathCount();
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder> 
-        getFamilyPathOrBuilderList();
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder(
-        int index);
-
-    // optional bool assign_seq_num = 2;
-    /**
-     * <code>optional bool assign_seq_num = 2;</code>
-     */
-    boolean hasAssignSeqNum();
-    /**
-     * <code>optional bool assign_seq_num = 2;</code>
-     */
-    boolean getAssignSeqNum();
-
-    // required .hbase.pb.DelegationToken fs_token = 3;
-    /**
-     * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-     */
-    boolean hasFsToken();
-    /**
-     * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-     */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken getFsToken();
-    /**
-     * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-     */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationTokenOrBuilder getFsTokenOrBuilder();
-
-    // required string bulk_token = 4;
-    /**
-     * <code>required string bulk_token = 4;</code>
-     */
-    boolean hasBulkToken();
-    /**
-     * <code>required string bulk_token = 4;</code>
-     */
-    java.lang.String getBulkToken();
-    /**
-     * <code>required string bulk_token = 4;</code>
-     */
-    com.google.protobuf.ByteString
-        getBulkTokenBytes();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.SecureBulkLoadHFilesRequest}
-   */
-  public static final class SecureBulkLoadHFilesRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements SecureBulkLoadHFilesRequestOrBuilder {
-    // Use SecureBulkLoadHFilesRequest.newBuilder() to construct.
-    private SecureBulkLoadHFilesRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private SecureBulkLoadHFilesRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final SecureBulkLoadHFilesRequest defaultInstance;
-    public static SecureBulkLoadHFilesRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public SecureBulkLoadHFilesRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private SecureBulkLoadHFilesRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-                familyPath_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath>();
-                mutable_bitField0_ |= 0x00000001;
-              }
-              familyPath_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.PARSER, extensionRegistry));
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000001;
-              assignSeqNum_ = input.readBool();
-              break;
-            }
-            case 26: {
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000002) == 0x00000002)) {
-                subBuilder = fsToken_.toBuilder();
-              }
-              fsToken_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(fsToken_);
-                fsToken_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000002;
-              break;
-            }
-            case 34: {
-              bitField0_ |= 0x00000004;
-              bulkToken_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-          familyPath_ = java.util.Collections.unmodifiableList(familyPath_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.class, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<SecureBulkLoadHFilesRequest> PARSER =
-        new com.google.protobuf.AbstractParser<SecureBulkLoadHFilesRequest>() {
-      public SecureBulkLoadHFilesRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new SecureBulkLoadHFilesRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<SecureBulkLoadHFilesRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;
-    public static final int FAMILY_PATH_FIELD_NUMBER = 1;
-    private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_;
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> getFamilyPathList() {
-      return familyPath_;
-    }
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder> 
-        getFamilyPathOrBuilderList() {
-      return familyPath_;
-    }
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    public int getFamilyPathCount() {
-      return familyPath_.size();
-    }
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index) {
-      return familyPath_.get(index);
-    }
-    /**
-     * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-     */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder(
-        int index) {
-      return familyPath_.get(index);
-    }
-
-    // optional bool assign_seq_num = 2;
-    public static final int ASSIGN_SEQ_NUM_FIELD_NUMBER = 2;
-    private boolean assignSeqNum_;
-    /**
-     * <code>optional bool assign_seq_num = 2;</code>
-     */
-    public boolean hasAssignSeqNum() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>optional bool assign_seq_num = 2;</code>
-     */
-    public boolean getAssignSeqNum() {
-      return assignSeqNum_;
-    }
-
-    // required .hbase.pb.DelegationToken fs_token = 3;
-    public static final int FS_TOKEN_FIELD_NUMBER = 3;
-    private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken fsToken_;
-    /**
-     * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-     */
-    public boolean hasFsToken() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-     */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken getFsToken() {
-      return fsToken_;
-    }
-    /**
-     * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-     */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationTokenOrBuilder getFsTokenOrBuilder() {
-      return fsToken_;
-    }
-
-    // required string bulk_token = 4;
-    public static final int BULK_TOKEN_FIELD_NUMBER = 4;
-    private java.lang.Object bulkToken_;
-    /**
-     * <code>required string bulk_token = 4;</code>
-     */
-    public boolean hasBulkToken() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>required string bulk_token = 4;</code>
-     */
-    public java.lang.String getBulkToken() {
-      java.lang.Object ref = bulkToken_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        com.google.protobuf.ByteString bs = 
-            (com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          bulkToken_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <code>required string bulk_token = 4;</code>
-     */
-    public com.google.protobuf.ByteString
-        getBulkTokenBytes() {
-      java.lang.Object ref = bulkToken_;
-      if (ref instanceof java.lang.String) {
-        com.google.protobuf.ByteString b = 
-            com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        bulkToken_ = b;
-        return b;
-      } else {
-        return (com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    private void initFields() {
-      familyPath_ = java.util.Collections.emptyList();
-      assignSeqNum_ = false;
-      fsToken_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.getDefaultInstance();
-      bulkToken_ = "";
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasFsToken()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!hasBulkToken()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      for (int i = 0; i < getFamilyPathCount(); i++) {
-        if (!getFamilyPath(i).isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      for (int i = 0; i < familyPath_.size(); i++) {
-        output.writeMessage(1, familyPath_.get(i));
-      }
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBool(2, assignSeqNum_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeMessage(3, fsToken_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeBytes(4, getBulkTokenBytes());
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      for (int i = 0; i < familyPath_.size(); i++) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, familyPath_.get(i));
-      }
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBoolSize(2, assignSeqNum_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(3, fsToken_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(4, getBulkTokenBytes());
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest other = (org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest) obj;
-
-      boolean result = true;
-      result = result && getFamilyPathList()
-          .equals(other.getFamilyPathList());
-      result = result && (hasAssignSeqNum() == other.hasAssignSeqNum());
-      if (hasAssignSeqNum()) {
-        result = result && (getAssignSeqNum()
-            == other.getAssignSeqNum());
-      }
-      result = result && (hasFsToken() == other.hasFsToken());
-      if (hasFsToken()) {
-        result = result && getFsToken()
-            .equals(other.getFsToken());
-      }
-      result = result && (hasBulkToken() == other.hasBulkToken());
-      if (hasBulkToken()) {
-        result = result && getBulkToken()
-            .equals(other.getBulkToken());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (getFamilyPathCount() > 0) {
-        hash = (37 * hash) + FAMILY_PATH_FIELD_NUMBER;
-        hash = (53 * hash) + getFamilyPathList().hashCode();
-      }
-      if (hasAssignSeqNum()) {
-        hash = (37 * hash) + ASSIGN_SEQ_NUM_FIELD_NUMBER;
-        hash = (53 * hash) + hashBoolean(getAssignSeqNum());
-      }
-      if (hasFsToken()) {
-        hash = (37 * hash) + FS_TOKEN_FIELD_NUMBER;
-        hash = (53 * hash) + getFsToken().hashCode();
-      }
-      if (hasBulkToken()) {
-        hash = (37 * hash) + BULK_TOKEN_FIELD_NUMBER;
-        hash = (53 * hash) + getBulkToken().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.SecureBulkLoadHFilesRequest}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.class, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-          getFamilyPathFieldBuilder();
-          getFsTokenFieldBuilder();
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        if (familyPathBuilder_ == null) {
-          familyPath_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000001);
-        } else {
-          familyPathBuilder_.clear();
-        }
-        assignSeqNum_ = false;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        if (fsTokenBuilder_ == null) {
-          fsToken_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.getDefaultInstance();
-        } else {
-          fsTokenBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000004);
-        bulkToken_ = "";
-        bitField0_ = (bitField0_ & ~0x00000008);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest build() {
-        org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest buildPartial() {
-        org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest result = new org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (familyPathBuilder_ == null) {
-          if (((bitField0_ & 0x00000001) == 0x00000001)) {
-            familyPath_ = java.util.Collections.unmodifiableList(familyPath_);
-            bitField0_ = (bitField0_ & ~0x00000001);
-          }
-          result.familyPath_ = familyPath_;
-        } else {
-          result.familyPath_ = familyPathBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.assignSeqNum_ = assignSeqNum_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        if (fsTokenBuilder_ == null) {
-          result.fsToken_ = fsToken_;
-        } else {
-          result.fsToken_ = fsTokenBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        result.bulkToken_ = bulkToken_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest other) {
-        if (other == org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.getDefaultInstance()) return this;
-        if (familyPathBuilder_ == null) {
-          if (!other.familyPath_.isEmpty()) {
-            if (familyPath_.isEmpty()) {
-              familyPath_ = other.familyPath_;
-              bitField0_ = (bitField0_ & ~0x00000001);
-            } else {
-              ensureFamilyPathIsMutable();
-              familyPath_.addAll(other.familyPath_);
-            }
-            onChanged();
-          }
-        } else {
-          if (!other.familyPath_.isEmpty()) {
-            if (familyPathBuilder_.isEmpty()) {
-              familyPathBuilder_.dispose();
-              familyPathBuilder_ = null;
-              familyPath_ = other.familyPath_;
-              bitField0_ = (bitField0_ & ~0x00000001);
-              familyPathBuilder_ = 
-                com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
-                   getFamilyPathFieldBuilder() : null;
-            } else {
-              familyPathBuilder_.addAllMessages(other.familyPath_);
-            }
-          }
-        }
-        if (other.hasAssignSeqNum()) {
-          setAssignSeqNum(other.getAssignSeqNum());
-        }
-        if (other.hasFsToken()) {
-          mergeFsToken(other.getFsToken());
-        }
-        if (other.hasBulkToken()) {
-          bitField0_ |= 0x00000008;
-          bulkToken_ = other.bulkToken_;
-          onChanged();
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasFsToken()) {
-          
-          return false;
-        }
-        if (!hasBulkToken()) {
-          
-          return false;
-        }
-        for (int i = 0; i < getFamilyPathCount(); i++) {
-          if (!getFamilyPath(i).isInitialized()) {
-            
-            return false;
-          }
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;
-      private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> familyPath_ =
-        java.util.Collections.emptyList();
-      private void ensureFamilyPathIsMutable() {
-        if (!((bitField0_ & 0x00000001) == 0x00000001)) {
-          familyPath_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath>(familyPath_);
-          bitField0_ |= 0x00000001;
-         }
-      }
-
-      private com.google.protobuf.RepeatedFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder> familyPathBuilder_;
-
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> getFamilyPathList() {
-        if (familyPathBuilder_ == null) {
-          return java.util.Collections.unmodifiableList(familyPath_);
-        } else {
-          return familyPathBuilder_.getMessageList();
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public int getFamilyPathCount() {
-        if (familyPathBuilder_ == null) {
-          return familyPath_.size();
-        } else {
-          return familyPathBuilder_.getCount();
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath getFamilyPath(int index) {
-        if (familyPathBuilder_ == null) {
-          return familyPath_.get(index);
-        } else {
-          return familyPathBuilder_.getMessage(index);
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder setFamilyPath(
-          int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath value) {
-        if (familyPathBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureFamilyPathIsMutable();
-          familyPath_.set(index, value);
-          onChanged();
-        } else {
-          familyPathBuilder_.setMessage(index, value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder setFamilyPath(
-          int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder builderForValue) {
-        if (familyPathBuilder_ == null) {
-          ensureFamilyPathIsMutable();
-          familyPath_.set(index, builderForValue.build());
-          onChanged();
-        } else {
-          familyPathBuilder_.setMessage(index, builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder addFamilyPath(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath value) {
-        if (familyPathBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureFamilyPathIsMutable();
-          familyPath_.add(value);
-          onChanged();
-        } else {
-          familyPathBuilder_.addMessage(value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder addFamilyPath(
-          int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath value) {
-        if (familyPathBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureFamilyPathIsMutable();
-          familyPath_.add(index, value);
-          onChanged();
-        } else {
-          familyPathBuilder_.addMessage(index, value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder addFamilyPath(
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder builderForValue) {
-        if (familyPathBuilder_ == null) {
-          ensureFamilyPathIsMutable();
-          familyPath_.add(builderForValue.build());
-          onChanged();
-        } else {
-          familyPathBuilder_.addMessage(builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder addFamilyPath(
-          int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder builderForValue) {
-        if (familyPathBuilder_ == null) {
-          ensureFamilyPathIsMutable();
-          familyPath_.add(index, builderForValue.build());
-          onChanged();
-        } else {
-          familyPathBuilder_.addMessage(index, builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder addAllFamilyPath(
-          java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath> values) {
-        if (familyPathBuilder_ == null) {
-          ensureFamilyPathIsMutable();
-          super.addAll(values, familyPath_);
-          onChanged();
-        } else {
-          familyPathBuilder_.addAllMessages(values);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder clearFamilyPath() {
-        if (familyPathBuilder_ == null) {
-          familyPath_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000001);
-          onChanged();
-        } else {
-          familyPathBuilder_.clear();
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public Builder removeFamilyPath(int index) {
-        if (familyPathBuilder_ == null) {
-          ensureFamilyPathIsMutable();
-          familyPath_.remove(index);
-          onChanged();
-        } else {
-          familyPathBuilder_.remove(index);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder getFamilyPathBuilder(
-          int index) {
-        return getFamilyPathFieldBuilder().getBuilder(index);
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder getFamilyPathOrBuilder(
-          int index) {
-        if (familyPathBuilder_ == null) {
-          return familyPath_.get(index);  } else {
-          return familyPathBuilder_.getMessageOrBuilder(index);
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder> 
-           getFamilyPathOrBuilderList() {
-        if (familyPathBuilder_ != null) {
-          return familyPathBuilder_.getMessageOrBuilderList();
-        } else {
-          return java.util.Collections.unmodifiableList(familyPath_);
-        }
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder addFamilyPathBuilder() {
-        return getFamilyPathFieldBuilder().addBuilder(
-            org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.getDefaultInstance());
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder addFamilyPathBuilder(
-          int index) {
-        return getFamilyPathFieldBuilder().addBuilder(
-            index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.getDefaultInstance());
-      }
-      /**
-       * <code>repeated .hbase.pb.BulkLoadHFileRequest.FamilyPath family_path = 1;</code>
-       */
-      public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder> 
-           getFamilyPathBuilderList() {
-        return getFamilyPathFieldBuilder().getBuilderList();
-      }
-      private com.google.protobuf.RepeatedFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder> 
-          getFamilyPathFieldBuilder() {
-        if (familyPathBuilder_ == null) {
-          familyPathBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPath.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest.FamilyPathOrBuilder>(
-                  familyPath_,
-                  ((bitField0_ & 0x00000001) == 0x00000001),
-                  getParentForChildren(),
-                  isClean());
-          familyPath_ = null;
-        }
-        return familyPathBuilder_;
-      }
-
-      // optional bool assign_seq_num = 2;
-      private boolean assignSeqNum_ ;
-      /**
-       * <code>optional bool assign_seq_num = 2;</code>
-       */
-      public boolean hasAssignSeqNum() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional bool assign_seq_num = 2;</code>
-       */
-      public boolean getAssignSeqNum() {
-        return assignSeqNum_;
-      }
-      /**
-       * <code>optional bool assign_seq_num = 2;</code>
-       */
-      public Builder setAssignSeqNum(boolean value) {
-        bitField0_ |= 0x00000002;
-        assignSeqNum_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bool assign_seq_num = 2;</code>
-       */
-      public Builder clearAssignSeqNum() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        assignSeqNum_ = false;
-        onChanged();
-        return this;
-      }
-
-      // required .hbase.pb.DelegationToken fs_token = 3;
-      private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken fsToken_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationTokenOrBuilder> fsTokenBuilder_;
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      public boolean hasFsToken() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
-      }
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken getFsToken() {
-        if (fsTokenBuilder_ == null) {
-          return fsToken_;
-        } else {
-          return fsTokenBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      public Builder setFsToken(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken value) {
-        if (fsTokenBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          fsToken_ = value;
-          onChanged();
-        } else {
-          fsTokenBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000004;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      public Builder setFsToken(
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.Builder builderForValue) {
-        if (fsTokenBuilder_ == null) {
-          fsToken_ = builderForValue.build();
-          onChanged();
-        } else {
-          fsTokenBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000004;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      public Builder mergeFsToken(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken value) {
-        if (fsTokenBuilder_ == null) {
-          if (((bitField0_ & 0x00000004) == 0x00000004) &&
-              fsToken_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.getDefaultInstance()) {
-            fsToken_ =
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.newBuilder(fsToken_).mergeFrom(value).buildPartial();
-          } else {
-            fsToken_ = value;
-          }
-          onChanged();
-        } else {
-          fsTokenBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000004;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      public Builder clearFsToken() {
-        if (fsTokenBuilder_ == null) {
-          fsToken_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.getDefaultInstance();
-          onChanged();
-        } else {
-          fsTokenBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000004);
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.Builder getFsTokenBuilder() {
-        bitField0_ |= 0x00000004;
-        onChanged();
-        return getFsTokenFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationTokenOrBuilder getFsTokenOrBuilder() {
-        if (fsTokenBuilder_ != null) {
-          return fsTokenBuilder_.getMessageOrBuilder();
-        } else {
-          return fsToken_;
-        }
-      }
-      /**
-       * <code>required .hbase.pb.DelegationToken fs_token = 3;</code>
-       */
-      private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationTokenOrBuilder> 
-          getFsTokenFieldBuilder() {
-        if (fsTokenBuilder_ == null) {
-          fsTokenBuilder_ = new com.google.protobuf.SingleFieldBuilder<
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationToken.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.DelegationTokenOrBuilder>(
-                  fsToken_,
-                  getParentForChildren(),
-                  isClean());
-          fsToken_ = null;
-        }
-        return fsTokenBuilder_;
-      }
-
-      // required string bulk_token = 4;
-      private java.lang.Object bulkToken_ = "";
-      /**
-       * <code>required string bulk_token = 4;</code>
-       */
-      public boolean hasBulkToken() {
-        return ((bitField0_ & 0x00000008) == 0x00000008);
-      }
-      /**
-       * <code>required string bulk_token = 4;</code>
-       */
-      public java.lang.String getBulkToken() {
-        java.lang.Object ref = bulkToken_;
-        if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          bulkToken_ = s;
-          return s;
-        } else {
-          return (java.lang.String) ref;
-        }
-      }
-      /**
-       * <code>required string bulk_token = 4;</code>
-       */
-      public com.google.protobuf.ByteString
-          getBulkTokenBytes() {
-        java.lang.Object ref = bulkToken_;
-        if (ref instanceof String) {
-          com.google.protobuf.ByteString b = 
-              com.google.protobuf.ByteString.copyFromUtf8(
-                  (java.lang.String) ref);
-          bulkToken_ = b;
-          return b;
-        } else {
-          return (com.google.protobuf.ByteString) ref;
-        }
-      }
-      /**
-       * <code>required string bulk_token = 4;</code>
-       */
-      public Builder setBulkToken(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000008;
-        bulkToken_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string bulk_token = 4;</code>
-       */
-      public Builder clearBulkToken() {
-        bitField0_ = (bitField0_ & ~0x00000008);
-        bulkToken_ = getDefaultInstance().getBulkToken();
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string bulk_token = 4;</code>
-       */
-      public Builder setBulkTokenBytes(
-          com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000008;
-        bulkToken_ = value;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.SecureBulkLoadHFilesRequest)
-    }
-
-    static {
-      defaultInstance = new SecureBulkLoadHFilesRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.SecureBulkLoadHFilesRequest)
-  }
-
-  public interface SecureBulkLoadHFilesResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required bool loaded = 1;
-    /**
-     * <code>required bool loaded = 1;</code>
-     */
-    boolean hasLoaded();
-    /**
-     * <code>required bool loaded = 1;</code>
-     */
-    boolean getLoaded();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.SecureBulkLoadHFilesResponse}
-   */
-  public static final class SecureBulkLoadHFilesResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements SecureBulkLoadHFilesResponseOrBuilder {
-    // Use SecureBulkLoadHFilesResponse.newBuilder() to construct.
-    private SecureBulkLoadHFilesResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private SecureBulkLoadHFilesResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final SecureBulkLoadHFilesResponse defaultInstance;
-    public static SecureBulkLoadHFilesResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public SecureBulkLoadHFilesResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private SecureBulkLoadHFilesResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              loaded_ = input.readBool();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.class, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<SecureBulkLoadHFilesResponse> PARSER =
-        new com.google.protobuf.AbstractParser<SecureBulkLoadHFilesResponse>() {
-      public SecureBulkLoadHFilesResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new SecureBulkLoadHFilesResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<SecureBulkLoadHFilesResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required bool loaded = 1;
-    public static final int LOADED_FIELD_NUMBER = 1;
-    private boolean loaded_;
-    /**
-     * <code>required bool loaded = 1;</code>
-     */
-    public boolean hasLoaded() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required bool loaded = 1;</code>
-     */
-    public boolean getLoaded() {
-      return loaded_;
-    }
-
-    private void initFields() {
-      loaded_ = false;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasLoaded()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBool(1, loaded_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBoolSize(1, loaded_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse other = (org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse) obj;
-
-      boolean result = true;
-      result = result && (hasLoaded() == other.hasLoaded());
-      if (hasLoaded()) {
-        result = result && (getLoaded()
-            == other.getLoaded());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasLoaded()) {
-        hash = (37 * hash) + LOADED_FIELD_NUMBER;
-        hash = (53 * hash) + hashBoolean(getLoaded());
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.SecureBulkLoadHFilesResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.class, org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        loaded_ = false;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.internal_static_hbase_pb_SecureBulkLoadHFilesResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse build() {
-        org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse buildPartial() {
-        org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse result = new org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.loaded_ = loaded_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse other) {
-        if (other == org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.getDefaultInstance()) return this;
-        if (other.hasLoaded()) {
-          setLoaded(other.getLoaded());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasLoaded()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required bool loaded = 1;
-      private boolean loaded_ ;
-      /**
-       * <code>required bool loaded = 1;</code>
-       */
-      public boolean hasLoaded() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required bool loaded = 1;</code>
-       */
-      public boolean getLoaded() {
-        return loaded_;
-      }
-      /**
-       * <code>required bool loaded = 1;</code>
-       */
-      public Builder setLoaded(boolean value) {
-        bitField0_ |= 0x00000001;
-        loaded_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required bool loaded = 1;</code>
-       */
-      public Builder clearLoaded() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        loaded_ = false;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.SecureBulkLoadHFilesResponse)
-    }
-
-    static {
-      defaultInstance = new SecureBulkLoadHFilesResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.SecureBulkLoadHFilesResponse)
-  }
-
-  /**
-   * Protobuf service {@code hbase.pb.SecureBulkLoadService}
-   */
-  public static abstract class SecureBulkLoadService
-      implements com.google.protobuf.Service {
-    protected SecureBulkLoadService() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc PrepareBulkLoad(.hbase.pb.PrepareBulkLoadRequest) returns (.hbase.pb.PrepareBulkLoadResponse);</code>
-       */
-      public abstract void prepareBulkLoad(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse> done);
-
-      /**
-       * <code>rpc SecureBulkLoadHFiles(.hbase.pb.SecureBulkLoadHFilesRequest) returns (.hbase.pb.SecureBulkLoadHFilesResponse);</code>
-       */
-      public abstract void secureBulkLoadHFiles(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse> done);
-
-      /**
-       * <code>rpc CleanupBulkLoad(.hbase.pb.CleanupBulkLoadRequest) returns (.hbase.pb.CleanupBulkLoadResponse);</code>
-       */
-      public abstract void cleanupBulkLoad(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse> done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new SecureBulkLoadService() {
-        @java.lang.Override
-        public  void prepareBulkLoad(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse> done) {
-          impl.prepareBulkLoad(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void secureBulkLoadHFiles(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse> done) {
-          impl.secureBulkLoadHFiles(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void cleanupBulkLoad(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse> done) {
-          impl.cleanupBulkLoad(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.prepareBulkLoad(controller, (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest)request);
-            case 1:
-              return impl.secureBulkLoadHFiles(controller, (org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest)request);
-            case 2:
-              return impl.cleanupBulkLoad(controller, (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.getDefaultInstance();
-            case 2:
-              return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.getDefaultInstance();
-            case 2:
-              return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc PrepareBulkLoad(.hbase.pb.PrepareBulkLoadRequest) returns (.hbase.pb.PrepareBulkLoadResponse);</code>
-     */
-    public abstract void prepareBulkLoad(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse> done);
-
-    /**
-     * <code>rpc SecureBulkLoadHFiles(.hbase.pb.SecureBulkLoadHFilesRequest) returns (.hbase.pb.SecureBulkLoadHFilesResponse);</code>
-     */
-    public abstract void secureBulkLoadHFiles(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse> done);
-
-    /**
-     * <code>rpc CleanupBulkLoad(.hbase.pb.CleanupBulkLoadRequest) returns (.hbase.pb.CleanupBulkLoadResponse);</code>
-     */
-    public abstract void cleanupBulkLoad(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse> done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.prepareBulkLoad(controller, (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse>specializeCallback(
-              done));
-          return;
-        case 1:
-          this.secureBulkLoadHFiles(controller, (org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse>specializeCallback(
-              done));
-          return;
-        case 2:
-          this.cleanupBulkLoad(controller, (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest.getDefaultInstance();
-        case 2:
-          return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.getDefaultInstance();
-        case 2:
-          return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadService implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void prepareBulkLoad(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse.class,
-            org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse.getDefaultInstance()));
-      }
-
-      public  void secureBulkLoadHFiles(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.class,
-            org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.getDefaultInstance()));
-      }
-
-      public  void cleanupBulkLoad(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(2),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse.class,
-            org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse prepareBulkLoad(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest request)
-          throws com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse secureBulkLoadHFiles(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest request)
-          throws com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse cleanupBulkLoad(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest request)
-          throws com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse prepareBulkLoad(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.PrepareBulkLoadResponse.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse secureBulkLoadHFiles(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.SecureBulkLoadProtos.SecureBulkLoadHFilesResponse.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse cleanupBulkLoad(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(2),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.CleanupBulkLoadResponse.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.SecureBulkLoadService)
-  }
-
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_hbase_pb_SecureBulkLoadHFilesRequest_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_hbase_pb_SecureBulkLoadHFilesRequest_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_hbase_pb_SecureBulkLoadHFilesResponse_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_hbase_pb_SecureBulkLoadHFilesResponse_fieldAccessorTable;
-
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\024SecureBulkLoad.proto\022\010hbase.pb\032\014Client" +
-      ".proto\"\266\001\n\033SecureBulkLoadHFilesRequest\022>" +
-      "\n\013family_path\030\001 \003(\0132).hbase.pb.BulkLoadH" +
-      "FileRequest.FamilyPath\022\026\n\016assign_seq_num" +
-      "\030\002 \001(\010\022+\n\010fs_token\030\003 \002(\0132\031.hbase.pb.Dele" +
-      "gationToken\022\022\n\nbulk_token\030\004 \002(\t\".\n\034Secur" +
-      "eBulkLoadHFilesResponse\022\016\n\006loaded\030\001 \002(\0102" +
-      "\256\002\n\025SecureBulkLoadService\022V\n\017PrepareBulk" +
-      "Load\022 .hbase.pb.PrepareBulkLoadRequest\032!" +
-      ".hbase.pb.PrepareBulkLoadResponse\022e\n\024Sec",
-      "ureBulkLoadHFiles\022%.hbase.pb.SecureBulkL" +
-      "oadHFilesRequest\032&.hbase.pb.SecureBulkLo" +
-      "adHFilesResponse\022V\n\017CleanupBulkLoad\022 .hb" +
-      "ase.pb.CleanupBulkLoadRequest\032!.hbase.pb" +
-      ".CleanupBulkLoadResponseBJ\n*org.apache.h" +
-      "adoop.hbase.protobuf.generatedB\024SecureBu" +
-      "lkLoadProtosH\001\210\001\001\240\001\001"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_hbase_pb_SecureBulkLoadHFilesRequest_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_hbase_pb_SecureBulkLoadHFilesRequest_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_pb_SecureBulkLoadHF

<TRUNCATED>

[21/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapEntry.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapEntry.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapEntry.java
deleted file mode 100644
index d586cc7..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapEntry.java
+++ /dev/null
@@ -1,449 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-
-import java.io.IOException;
-import java.util.Collections;
-import java.util.Map;
-import java.util.TreeMap;
-
-/**
- * Implements MapEntry messages.
- *
- * In reflection API, map fields will be treated as repeated message fields and
- * each map entry is accessed as a message. This MapEntry class is used to
- * represent these map entry messages in reflection API.
- *
- * Protobuf internal. Users shouldn't use this class.
- */
-public final class MapEntry<K, V> extends AbstractMessage {
-
-  private static final class Metadata<K, V> extends MapEntryLite.Metadata<K, V> {
-
-    public final Descriptor descriptor;
-    public final Parser<MapEntry<K, V>> parser;
-
-    public Metadata(
-        Descriptor descriptor,
-        MapEntry<K, V> defaultInstance,
-        WireFormat.FieldType keyType,
-        WireFormat.FieldType valueType) {
-      super(keyType, defaultInstance.key, valueType, defaultInstance.value);
-      this.descriptor = descriptor;
-      this.parser = new AbstractParser<MapEntry<K, V>>() {
-
-        @Override
-        public MapEntry<K, V> parsePartialFrom(
-            CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-            throws InvalidProtocolBufferException {
-          return new MapEntry<K, V>(Metadata.this, input, extensionRegistry);
-        }
-      };
-    }
-  }
-
-  private final K key;
-  private final V value;
-  private final Metadata<K, V> metadata;
-
-  /** Create a default MapEntry instance. */
-  private MapEntry(
-      Descriptor descriptor,
-      WireFormat.FieldType keyType, K defaultKey,
-      WireFormat.FieldType valueType, V defaultValue) {
-    this.key = defaultKey;
-    this.value = defaultValue;
-    this.metadata = new Metadata<K, V>(descriptor, this, keyType, valueType);
-  }
-
-  /** Create a MapEntry with the provided key and value. */
-  private MapEntry(Metadata metadata, K key, V value) {
-    this.key = key;
-    this.value = value;
-    this.metadata = metadata;
-  }
-
-  /** Parsing constructor. */
-  private MapEntry(
-      Metadata<K, V> metadata,
-      CodedInputStream input,
-      ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    try {
-      this.metadata = metadata;
-      Map.Entry<K, V> entry = MapEntryLite.parseEntry(input, metadata, extensionRegistry);
-      this.key = entry.getKey();
-      this.value = entry.getValue();
-    } catch (InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (IOException e) {
-      throw new InvalidProtocolBufferException(e).setUnfinishedMessage(this);
-    }
-  }
-
-  /**
-   * Create a default MapEntry instance. A default MapEntry instance should be
-   * created only once for each map entry message type. Generated code should
-   * store the created default instance and use it later to create new MapEntry
-   * messages of the same type.
-   */
-  public static <K, V> MapEntry<K, V> newDefaultInstance(
-      Descriptor descriptor,
-      WireFormat.FieldType keyType, K defaultKey,
-      WireFormat.FieldType valueType, V defaultValue) {
-    return new MapEntry<K, V>(
-        descriptor, keyType, defaultKey, valueType, defaultValue);
-  }
-
-  public K getKey() {
-    return key;
-  }
-
-  public V getValue() {
-    return value;
-  }
-
-  private volatile int cachedSerializedSize = -1;
-
-  @Override
-  public int getSerializedSize() {
-    if (cachedSerializedSize != -1) {
-      return cachedSerializedSize;
-    }
-
-    int size = MapEntryLite.computeSerializedSize(metadata, key, value);
-    cachedSerializedSize = size;
-    return size;
-  }
-
-  @Override
-  public void writeTo(CodedOutputStream output) throws IOException {
-    MapEntryLite.writeTo(output, metadata, key, value);
-  }
-
-  @Override
-  public boolean isInitialized() {
-    return isInitialized(metadata, value);
-  }
-
-  @Override
-  public Parser<MapEntry<K, V>> getParserForType() {
-    return metadata.parser;
-  }
-
-  @Override
-  public Builder<K, V> newBuilderForType() {
-    return new Builder<K, V>(metadata);
-  }
-
-  @Override
-  public Builder<K, V> toBuilder() {
-    return new Builder<K, V>(metadata, key, value);
-  }
-
-  @Override
-  public MapEntry<K, V> getDefaultInstanceForType() {
-    return new MapEntry<K, V>(metadata, metadata.defaultKey, metadata.defaultValue);
-  }
-
-  @Override
-  public Descriptor getDescriptorForType() {
-    return metadata.descriptor;
-  }
-
-  @Override
-  public Map<FieldDescriptor, Object> getAllFields() {
-    TreeMap<FieldDescriptor, Object> result = new TreeMap<FieldDescriptor, Object>();
-    for (final FieldDescriptor field : metadata.descriptor.getFields()) {
-      if (hasField(field)) {
-        result.put(field, getField(field));
-      }
-    }
-    return Collections.unmodifiableMap(result);
-  }
-
-  private void checkFieldDescriptor(FieldDescriptor field) {
-    if (field.getContainingType() != metadata.descriptor) {
-      throw new RuntimeException(
-          "Wrong FieldDescriptor \"" + field.getFullName()
-          + "\" used in message \"" + metadata.descriptor.getFullName());
-    }
-  }
-
-  @Override
-  public boolean hasField(FieldDescriptor field) {
-    checkFieldDescriptor(field);;
-    // A MapEntry always contains two fields.
-    return true;
-  }
-
-  @Override
-  public Object getField(FieldDescriptor field) {
-    checkFieldDescriptor(field);
-    Object result = field.getNumber() == 1 ? getKey() : getValue();
-    // Convert enums to EnumValueDescriptor.
-    if (field.getType() == FieldDescriptor.Type.ENUM) {
-      result = field.getEnumType().findValueByNumberCreatingIfUnknown(
-          (java.lang.Integer) result);
-    }
-    return result;
-  }
-
-  @Override
-  public int getRepeatedFieldCount(FieldDescriptor field) {
-    throw new RuntimeException(
-        "There is no repeated field in a map entry message.");
-  }
-
-  @Override
-  public Object getRepeatedField(FieldDescriptor field, int index) {
-    throw new RuntimeException(
-        "There is no repeated field in a map entry message.");
-  }
-
-  @Override
-  public UnknownFieldSet getUnknownFields() {
-    return UnknownFieldSet.getDefaultInstance();
-  }
-
-  /**
-   * Builder to create {@link MapEntry} messages.
-   */
-  public static class Builder<K, V>
-      extends AbstractMessage.Builder<Builder<K, V>> {
-    private final Metadata<K, V> metadata;
-    private K key;
-    private V value;
-
-    private Builder(Metadata<K, V> metadata) {
-      this(metadata, metadata.defaultKey, metadata.defaultValue);
-    }
-
-    private Builder(Metadata<K, V> metadata, K key, V value) {
-      this.metadata = metadata;
-      this.key = key;
-      this.value = value;
-    }
-
-    public K getKey() {
-      return key;
-    }
-
-    public V getValue() {
-      return value;
-    }
-
-    public Builder<K, V> setKey(K key) {
-      this.key = key;
-      return this;
-    }
-
-    public Builder<K, V> clearKey() {
-      this.key = metadata.defaultKey;
-      return this;
-    }
-
-    public Builder<K, V> setValue(V value) {
-      this.value = value;
-      return this;
-    }
-
-    public Builder<K, V> clearValue() {
-      this.value = metadata.defaultValue;
-      return this;
-    }
-
-    @Override
-    public MapEntry<K, V> build() {
-      MapEntry<K, V> result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    @Override
-    public MapEntry<K, V> buildPartial() {
-      return new MapEntry<K, V>(metadata, key, value);
-    }
-
-    @Override
-    public Descriptor getDescriptorForType() {
-      return metadata.descriptor;
-    }
-
-    private void checkFieldDescriptor(FieldDescriptor field) {
-      if (field.getContainingType() != metadata.descriptor) {
-        throw new RuntimeException(
-            "Wrong FieldDescriptor \"" + field.getFullName()
-            + "\" used in message \"" + metadata.descriptor.getFullName());
-      }
-    }
-
-    @Override
-    public Message.Builder newBuilderForField(FieldDescriptor field) {
-      checkFieldDescriptor(field);;
-      // This method should be called for message fields and in a MapEntry
-      // message only the value field can possibly be a message field.
-      if (field.getNumber() != 2
-          || field.getJavaType() != FieldDescriptor.JavaType.MESSAGE) {
-        throw new RuntimeException(
-            "\"" + field.getFullName() + "\" is not a message value field.");
-      }
-      return ((Message) value).newBuilderForType();
-    }
-
-    @SuppressWarnings("unchecked")
-    @Override
-    public Builder<K, V> setField(FieldDescriptor field, Object value) {
-      checkFieldDescriptor(field);
-      if (field.getNumber() == 1) {
-        setKey((K) value);
-      } else {
-        if (field.getType() == FieldDescriptor.Type.ENUM) {
-          value = ((EnumValueDescriptor) value).getNumber();
-        } else if (field.getType() == FieldDescriptor.Type.MESSAGE) {
-          if (value != null && !metadata.defaultValue.getClass().isInstance(value)) {
-            // The value is not the exact right message type.  However, if it
-            // is an alternative implementation of the same type -- e.g. a
-            // DynamicMessage -- we should accept it.  In this case we can make
-            // a copy of the message.
-            value =
-                ((Message) metadata.defaultValue).toBuilder().mergeFrom((Message) value).build();
-          }
-        }
-        setValue((V) value);
-      }
-      return this;
-    }
-
-    @Override
-    public Builder<K, V> clearField(FieldDescriptor field) {
-      checkFieldDescriptor(field);
-      if (field.getNumber() == 1) {
-        clearKey();
-      } else {
-        clearValue();
-      }
-      return this;
-    }
-
-    @Override
-    public Builder<K, V> setRepeatedField(FieldDescriptor field, int index,
-        Object value) {
-      throw new RuntimeException(
-          "There is no repeated field in a map entry message.");
-    }
-
-    @Override
-    public Builder<K, V> addRepeatedField(FieldDescriptor field, Object value) {
-      throw new RuntimeException(
-          "There is no repeated field in a map entry message.");
-    }
-
-    @Override
-    public Builder<K, V> setUnknownFields(UnknownFieldSet unknownFields) {
-      // Unknown fields are discarded for MapEntry message.
-      return this;
-    }
-
-    @Override
-    public MapEntry<K, V> getDefaultInstanceForType() {
-      return new MapEntry<K, V>(metadata, metadata.defaultKey, metadata.defaultValue);
-    }
-
-    @Override
-    public boolean isInitialized() {
-      return MapEntry.isInitialized(metadata, value);
-    }
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFields() {
-      final TreeMap<FieldDescriptor, Object> result = new TreeMap<FieldDescriptor, Object>();
-      for (final FieldDescriptor field : metadata.descriptor.getFields()) {
-        if (hasField(field)) {
-          result.put(field, getField(field));
-        }
-      }
-      return Collections.unmodifiableMap(result);
-    }
-
-    @Override
-    public boolean hasField(FieldDescriptor field) {
-      checkFieldDescriptor(field);
-      return true;
-    }
-
-    @Override
-    public Object getField(FieldDescriptor field) {
-      checkFieldDescriptor(field);
-      Object result = field.getNumber() == 1 ? getKey() : getValue();
-      // Convert enums to EnumValueDescriptor.
-      if (field.getType() == FieldDescriptor.Type.ENUM) {
-        result = field.getEnumType().findValueByNumberCreatingIfUnknown((Integer) result);
-      }
-      return result;
-    }
-
-    @Override
-    public int getRepeatedFieldCount(FieldDescriptor field) {
-      throw new RuntimeException(
-          "There is no repeated field in a map entry message.");
-    }
-
-    @Override
-    public Object getRepeatedField(FieldDescriptor field, int index) {
-      throw new RuntimeException(
-          "There is no repeated field in a map entry message.");
-    }
-
-    @Override
-    public UnknownFieldSet getUnknownFields() {
-      return UnknownFieldSet.getDefaultInstance();
-    }
-
-    @Override
-    public Builder<K, V> clone() {
-      return new Builder(metadata, key, value);
-    }
-  }
-
-  private static <V> boolean isInitialized(Metadata metadata, V value) {
-    if (metadata.valueType.getJavaType() == WireFormat.JavaType.MESSAGE) {
-      return ((MessageLite) value).isInitialized();
-    }
-    return true;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapEntryLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapEntryLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapEntryLite.java
deleted file mode 100644
index d8a19bb..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapEntryLite.java
+++ /dev/null
@@ -1,226 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.util.AbstractMap;
-import java.util.Map;
-
-/**
- * Implements the lite version of map entry messages.
- *
- * This class serves as an utility class to help do serialization/parsing of
- * map entries. It's used in generated code and also in the full version
- * MapEntry message.
- *
- * Protobuf internal. Users shouldn't use.
- */
-public class MapEntryLite<K, V> {
-
-  static class Metadata<K, V> {
-    public final WireFormat.FieldType keyType;
-    public final K defaultKey;
-    public final WireFormat.FieldType valueType;
-    public final V defaultValue;
-
-    public Metadata(
-        WireFormat.FieldType keyType, K defaultKey,
-        WireFormat.FieldType valueType, V defaultValue) {
-      this.keyType = keyType;
-      this.defaultKey = defaultKey;
-      this.valueType = valueType;
-      this.defaultValue = defaultValue;
-    }
-  }
-
-  private static final int KEY_FIELD_NUMBER = 1;
-  private static final int VALUE_FIELD_NUMBER = 2;
-
-  private final Metadata<K, V> metadata;
-  private final K key;
-  private final V value;
-
-  /** Creates a default MapEntryLite message instance. */
-  private MapEntryLite(
-      WireFormat.FieldType keyType, K defaultKey,
-      WireFormat.FieldType valueType, V defaultValue) {
-    this.metadata = new Metadata<K, V>(keyType, defaultKey, valueType, defaultValue);
-    this.key = defaultKey;
-    this.value = defaultValue;
-  }
-
-  /** Creates a new MapEntryLite message. */
-  private MapEntryLite(Metadata<K, V> metadata, K key, V value) {
-    this.metadata = metadata;
-    this.key = key;
-    this.value = value;
-  }
-
-  public K getKey() {
-    return key;
-  }
-
-  public V getValue() {
-    return value;
-  }
-
-  /**
-   * Creates a default MapEntryLite message instance.
-   *
-   * This method is used by generated code to create the default instance for
-   * a map entry message. The created default instance should be used to create
-   * new map entry messages of the same type. For each map entry message, only
-   * one default instance should be created.
-   */
-  public static <K, V> MapEntryLite<K, V> newDefaultInstance(
-      WireFormat.FieldType keyType, K defaultKey,
-      WireFormat.FieldType valueType, V defaultValue) {
-    return new MapEntryLite<K, V>(
-        keyType, defaultKey, valueType, defaultValue);
-  }
-
-  static <K, V> void writeTo(CodedOutputStream output, Metadata<K, V> metadata, K key, V value)
-      throws IOException {
-    FieldSet.writeElement(output, metadata.keyType, KEY_FIELD_NUMBER, key);
-    FieldSet.writeElement(output, metadata.valueType, VALUE_FIELD_NUMBER, value);
-  }
-
-  static <K, V> int computeSerializedSize(Metadata<K, V> metadata, K key, V value) {
-    return FieldSet.computeElementSize(metadata.keyType, KEY_FIELD_NUMBER, key)
-        + FieldSet.computeElementSize(metadata.valueType, VALUE_FIELD_NUMBER, value);
-  }
-
-  @SuppressWarnings("unchecked")
-  static <T> T parseField(
-      CodedInputStream input, ExtensionRegistryLite extensionRegistry,
-      WireFormat.FieldType type, T value) throws IOException {
-    switch (type) {
-      case MESSAGE:
-        MessageLite.Builder subBuilder = ((MessageLite) value).toBuilder();
-        input.readMessage(subBuilder, extensionRegistry);
-        return (T) subBuilder.buildPartial();
-      case ENUM:
-        return (T) (java.lang.Integer) input.readEnum();
-      case GROUP:
-        throw new RuntimeException("Groups are not allowed in maps.");
-      default:
-        return (T) FieldSet.readPrimitiveField(input, type, true);
-    }
-  }
-
-  /**
-   * Serializes the provided key and value as though they were wrapped by a {@link MapEntryLite}
-   * to the output stream. This helper method avoids allocation of a {@link MapEntryLite}
-   * built with a key and value and is called from generated code directly.
-   */
-  public void serializeTo(CodedOutputStream output, int fieldNumber, K key, V value)
-      throws IOException {
-    output.writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-    output.writeUInt32NoTag(computeSerializedSize(metadata, key, value));
-    writeTo(output, metadata, key, value);
-  }
-
-  /**
-   * Computes the message size for the provided key and value as though they were wrapped
-   * by a {@link MapEntryLite}. This helper method avoids allocation of a {@link MapEntryLite}
-   * built with a key and value and is called from generated code directly.
-   */
-  public int computeMessageSize(int fieldNumber, K key, V value) {
-    return CodedOutputStream.computeTagSize(fieldNumber)
-        + CodedOutputStream.computeLengthDelimitedFieldSize(
-            computeSerializedSize(metadata, key, value));
-  }
-
-  /**
-   * Parses an entry off of the input as a {@link Map.Entry}. This helper requires an allocation
-   * so using {@link #parseInto} is preferred if possible.
-   */
-  public Map.Entry<K, V> parseEntry(ByteString bytes, ExtensionRegistryLite extensionRegistry)
-      throws IOException {
-    return parseEntry(bytes.newCodedInput(), metadata, extensionRegistry);
-  }
-
-  static <K, V> Map.Entry<K, V> parseEntry(
-      CodedInputStream input, Metadata<K, V> metadata, ExtensionRegistryLite extensionRegistry)
-          throws IOException{
-    K key = metadata.defaultKey;
-    V value = metadata.defaultValue;
-    while (true) {
-      int tag = input.readTag();
-      if (tag == 0) {
-        break;
-      }
-      if (tag == WireFormat.makeTag(KEY_FIELD_NUMBER, metadata.keyType.getWireType())) {
-        key = parseField(input, extensionRegistry, metadata.keyType, key);
-      } else if (tag == WireFormat.makeTag(VALUE_FIELD_NUMBER, metadata.valueType.getWireType())) {
-        value = parseField(input, extensionRegistry, metadata.valueType, value);
-      } else {
-        if (!input.skipField(tag)) {
-          break;
-        }
-      }
-    }
-    return new AbstractMap.SimpleImmutableEntry<K, V>(key, value);
-  }
-
-  /**
-   * Parses an entry off of the input into the map. This helper avoids allocaton of a
-   * {@link MapEntryLite} by parsing directly into the provided {@link MapFieldLite}.
-   */
-  public void parseInto(
-      MapFieldLite<K, V> map, CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-          throws IOException {
-    int length = input.readRawVarint32();
-    final int oldLimit = input.pushLimit(length);
-    K key = metadata.defaultKey;
-    V value = metadata.defaultValue;
-
-    while (true) {
-      int tag = input.readTag();
-      if (tag == 0) {
-        break;
-      }
-      if (tag == WireFormat.makeTag(KEY_FIELD_NUMBER, metadata.keyType.getWireType())) {
-        key = parseField(input, extensionRegistry, metadata.keyType, key);
-      } else if (tag == WireFormat.makeTag(VALUE_FIELD_NUMBER, metadata.valueType.getWireType())) {
-        value = parseField(input, extensionRegistry, metadata.valueType, value);
-      } else {
-        if (!input.skipField(tag)) {
-          break;
-        }
-      }
-    }
-
-    input.checkLastTagWas(0);
-    input.popLimit(oldLimit);
-    map.put(key, value);
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapField.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapField.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapField.java
deleted file mode 100644
index 42905d2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapField.java
+++ /dev/null
@@ -1,624 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.ArrayList;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.LinkedHashMap;
-import java.util.List;
-import java.util.Map;
-import java.util.Set;
-
-/**
- * Internal representation of map fields in generated messages.
- *
- * This class supports accessing the map field as a {@link Map} to be used in
- * generated API and also supports accessing the field as a {@link List} to be
- * used in reflection API. It keeps track of where the data is currently stored
- * and do necessary conversions between map and list.
- *
- * This class is a protobuf implementation detail. Users shouldn't use this
- * class directly.
- *
- * THREAD-SAFETY NOTE: Read-only access is thread-safe. Users can call getMap()
- * and getList() concurrently in multiple threads. If write-access is needed,
- * all access must be synchronized.
- */
-public class MapField<K, V> implements MutabilityOracle {
-  /**
-   * Indicates where the data of this map field is currently stored.
-   *
-   * MAP: Data is stored in mapData.
-   * LIST: Data is stored in listData.
-   * BOTH: mapData and listData have the same data.
-   *
-   * When the map field is accessed (through generated API or reflection API),
-   * it will shift between these 3 modes:
-   *
-   *          getMap()   getList()   getMutableMap()   getMutableList()
-   *   MAP      MAP        BOTH          MAP               LIST
-   *   LIST     BOTH       LIST          MAP               LIST
-   *   BOTH     BOTH       BOTH          MAP               LIST
-   *
-   * As the map field changes its mode, the list/map reference returned in a
-   * previous method call may be invalidated.
-   */
-  private enum StorageMode {MAP, LIST, BOTH}
-
-  private volatile boolean isMutable;
-  private volatile StorageMode mode;
-  private MutatabilityAwareMap<K, V> mapData;
-  private List<Message> listData;
-
-  // Convert between a map entry Message and a key-value pair.
-  private static interface Converter<K, V> {
-    Message convertKeyAndValueToMessage(K key, V value);
-    void convertMessageToKeyAndValue(Message message, Map<K, V> map);
-
-    Message getMessageDefaultInstance();
-  }
-
-  private static class ImmutableMessageConverter<K, V> implements Converter<K, V> {
-    private final MapEntry<K, V> defaultEntry;
-    public ImmutableMessageConverter(MapEntry<K, V> defaultEntry) {
-      this.defaultEntry = defaultEntry;
-    }
-
-    @Override
-    public Message convertKeyAndValueToMessage(K key, V value) {
-      return defaultEntry.newBuilderForType().setKey(key).setValue(value).buildPartial();
-    }
-
-    @Override
-    public void convertMessageToKeyAndValue(Message message, Map<K, V> map) {
-      MapEntry<K, V> entry = (MapEntry<K, V>) message;
-      map.put(entry.getKey(), entry.getValue());
-    }
-
-    @Override
-    public Message getMessageDefaultInstance() {
-      return defaultEntry;
-    }
-  }
-
-
-  private final Converter<K, V> converter;
-
-  private MapField(
-      Converter<K, V> converter,
-      StorageMode mode,
-      Map<K, V> mapData) {
-    this.converter = converter;
-    this.isMutable = true;
-    this.mode = mode;
-    this.mapData = new MutatabilityAwareMap<K, V>(this, mapData);
-    this.listData = null;
-  }
-
-  private MapField(
-      MapEntry<K, V> defaultEntry,
-      StorageMode mode,
-      Map<K, V> mapData) {
-    this(new ImmutableMessageConverter<K, V>(defaultEntry), mode, mapData);
-  }
-
-
-  /** Returns an immutable empty MapField. */
-  public static <K, V> MapField<K, V> emptyMapField(
-      MapEntry<K, V> defaultEntry) {
-    return new MapField<K, V>(
-        defaultEntry, StorageMode.MAP, Collections.<K, V>emptyMap());
-  }
-
-
-  /** Creates a new mutable empty MapField. */
-  public static <K, V> MapField<K, V> newMapField(MapEntry<K, V> defaultEntry) {
-    return new MapField<K, V>(
-        defaultEntry, StorageMode.MAP, new LinkedHashMap<K, V>());
-  }
-
-
-  private Message convertKeyAndValueToMessage(K key, V value) {
-    return converter.convertKeyAndValueToMessage(key, value);
-  }
-
-  @SuppressWarnings("unchecked")
-  private void convertMessageToKeyAndValue(Message message, Map<K, V> map) {
-    converter.convertMessageToKeyAndValue(message, map);
-  }
-
-  private List<Message> convertMapToList(MutatabilityAwareMap<K, V> mapData) {
-    List<Message> listData = new ArrayList<Message>();
-    for (Map.Entry<K, V> entry : mapData.entrySet()) {
-      listData.add(
-          convertKeyAndValueToMessage(
-              entry.getKey(), entry.getValue()));
-    }
-    return listData;
-  }
-
-  private MutatabilityAwareMap<K, V> convertListToMap(List<Message> listData) {
-    Map<K, V> mapData = new LinkedHashMap<K, V>();
-    for (Message item : listData) {
-      convertMessageToKeyAndValue(item, mapData);
-    }
-    return new MutatabilityAwareMap<K, V>(this, mapData);
-  }
-
-  /** Returns the content of this MapField as a read-only Map. */
-  public Map<K, V> getMap() {
-    if (mode == StorageMode.LIST) {
-      synchronized (this) {
-        if (mode == StorageMode.LIST) {
-          mapData = convertListToMap(listData);
-          mode = StorageMode.BOTH;
-        }
-      }
-    }
-    return Collections.unmodifiableMap(mapData);
-  }
-
-  /** Gets a mutable Map view of this MapField. */
-  public Map<K, V> getMutableMap() {
-    if (mode != StorageMode.MAP) {
-      if (mode == StorageMode.LIST) {
-        mapData = convertListToMap(listData);
-      }
-      listData = null;
-      mode = StorageMode.MAP;
-    }
-    return mapData;
-  }
-
-  public void mergeFrom(MapField<K, V> other) {
-    getMutableMap().putAll(MapFieldLite.copy(other.getMap()));
-  }
-
-  public void clear() {
-    mapData = new MutatabilityAwareMap<K, V>(this, new LinkedHashMap<K, V>());
-    mode = StorageMode.MAP;
-  }
-
-  @SuppressWarnings("unchecked")
-  @Override
-  public boolean equals(Object object) {
-    if (!(object instanceof MapField)) {
-      return false;
-    }
-    MapField<K, V> other = (MapField<K, V>) object;
-    return MapFieldLite.<K, V>equals(getMap(), other.getMap());
-  }
-
-  @Override
-  public int hashCode() {
-    return MapFieldLite.<K, V>calculateHashCodeForMap(getMap());
-  }
-
-  /** Returns a deep copy of this MapField. */
-  public MapField<K, V> copy() {
-    return new MapField<K, V>(
-        converter, StorageMode.MAP, MapFieldLite.copy(getMap()));
-  }
-
-  /** Gets the content of this MapField as a read-only List. */
-  List<Message> getList() {
-    if (mode == StorageMode.MAP) {
-      synchronized (this) {
-        if (mode == StorageMode.MAP) {
-          listData = convertMapToList(mapData);
-          mode = StorageMode.BOTH;
-        }
-      }
-    }
-    return Collections.unmodifiableList(listData);
-  }
-
-  /** Gets a mutable List view of this MapField. */
-  List<Message> getMutableList() {
-    if (mode != StorageMode.LIST) {
-      if (mode == StorageMode.MAP) {
-        listData = convertMapToList(mapData);
-      }
-      mapData = null;
-      mode = StorageMode.LIST;
-    }
-    return listData;
-  }
-
-  /**
-   * Gets the default instance of the message stored in the list view of this
-   * map field.
-   */
-  Message getMapEntryMessageDefaultInstance() {
-    return converter.getMessageDefaultInstance();
-  }
-
-  /**
-   * Makes this list immutable. All subsequent modifications will throw an
-   * {@link UnsupportedOperationException}.
-   */
-  public void makeImmutable() {
-    isMutable = false;
-  }
-
-  /**
-   * Returns whether this field can be modified.
-   */
-  public boolean isMutable() {
-    return isMutable;
-  }
-
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.hbase.shaded.com.google.protobuf.MutabilityOracle#ensureMutable()
-   */
-  @Override
-  public void ensureMutable() {
-    if (!isMutable()) {
-      throw new UnsupportedOperationException();
-    }
-  }
-
-  /**
-   * An internal map that checks for mutability before delegating.
-   */
-  private static class MutatabilityAwareMap<K, V> implements Map<K, V> {
-    private final MutabilityOracle mutabilityOracle;
-    private final Map<K, V> delegate;
-
-    MutatabilityAwareMap(MutabilityOracle mutabilityOracle, Map<K, V> delegate) {
-      this.mutabilityOracle = mutabilityOracle;
-      this.delegate = delegate;
-    }
-
-    @Override
-    public int size() {
-      return delegate.size();
-    }
-
-    @Override
-    public boolean isEmpty() {
-      return delegate.isEmpty();
-    }
-
-    @Override
-    public boolean containsKey(Object key) {
-      return delegate.containsKey(key);
-    }
-
-    @Override
-    public boolean containsValue(Object value) {
-      return delegate.containsValue(value);
-    }
-
-    @Override
-    public V get(Object key) {
-      return delegate.get(key);
-    }
-
-    @Override
-    public V put(K key, V value) {
-      mutabilityOracle.ensureMutable();
-      return delegate.put(key, value);
-    }
-
-    @Override
-    public V remove(Object key) {
-      mutabilityOracle.ensureMutable();
-      return delegate.remove(key);
-    }
-
-    @Override
-    public void putAll(Map<? extends K, ? extends V> m) {
-      mutabilityOracle.ensureMutable();
-      delegate.putAll(m);
-    }
-
-    @Override
-    public void clear() {
-      mutabilityOracle.ensureMutable();
-      delegate.clear();
-    }
-
-    @Override
-    public Set<K> keySet() {
-      return new MutatabilityAwareSet<K>(mutabilityOracle, delegate.keySet());
-    }
-
-    @Override
-    public Collection<V> values() {
-      return new MutatabilityAwareCollection<V>(mutabilityOracle, delegate.values());
-    }
-
-    @Override
-    public Set<java.util.Map.Entry<K, V>> entrySet() {
-      return new MutatabilityAwareSet<Entry<K, V>>(mutabilityOracle, delegate.entrySet());
-    }
-
-    @Override
-    public boolean equals(Object o) {
-      return delegate.equals(o);
-    }
-
-    @Override
-    public int hashCode() {
-      return delegate.hashCode();
-    }
-
-    @Override
-    public String toString() {
-      return delegate.toString();
-    }
-
-    /**
-     * An internal collection that checks for mutability before delegating.
-     */
-    private static class MutatabilityAwareCollection<E> implements Collection<E> {
-      private final MutabilityOracle mutabilityOracle;
-      private final Collection<E> delegate;
-
-      MutatabilityAwareCollection(MutabilityOracle mutabilityOracle, Collection<E> delegate) {
-        this.mutabilityOracle = mutabilityOracle;
-        this.delegate = delegate;
-      }
-
-      @Override
-      public int size() {
-        return delegate.size();
-      }
-
-      @Override
-      public boolean isEmpty() {
-        return delegate.isEmpty();
-      }
-
-      @Override
-      public boolean contains(Object o) {
-        return delegate.contains(o);
-      }
-
-      @Override
-      public Iterator<E> iterator() {
-        return new MutatabilityAwareIterator<E>(mutabilityOracle, delegate.iterator());
-      }
-
-      @Override
-      public Object[] toArray() {
-        return delegate.toArray();
-      }
-
-      @Override
-      public <T> T[] toArray(T[] a) {
-        return delegate.toArray(a);
-      }
-
-      @Override
-      public boolean add(E e) {
-        // Unsupported operation in the delegate.
-        throw new UnsupportedOperationException();
-      }
-
-      @Override
-      public boolean remove(Object o) {
-        mutabilityOracle.ensureMutable();
-        return delegate.remove(o);
-      }
-
-      @Override
-      public boolean containsAll(Collection<?> c) {
-        return delegate.containsAll(c);
-      }
-
-      @Override
-      public boolean addAll(Collection<? extends E> c) {
-        // Unsupported operation in the delegate.
-        throw new UnsupportedOperationException();
-      }
-
-      @Override
-      public boolean removeAll(Collection<?> c) {
-        mutabilityOracle.ensureMutable();
-        return delegate.removeAll(c);
-      }
-
-      @Override
-      public boolean retainAll(Collection<?> c) {
-        mutabilityOracle.ensureMutable();
-        return delegate.retainAll(c);
-      }
-
-      @Override
-      public void clear() {
-        mutabilityOracle.ensureMutable();
-        delegate.clear();
-      }
-
-      @Override
-      public boolean equals(Object o) {
-        return delegate.equals(o);
-      }
-
-      @Override
-      public int hashCode() {
-        return delegate.hashCode();
-      }
-
-      @Override
-      public String toString() {
-        return delegate.toString();
-      }
-    }
-
-    /**
-     * An internal set that checks for mutability before delegating.
-     */
-    private static class MutatabilityAwareSet<E> implements Set<E> {
-      private final MutabilityOracle mutabilityOracle;
-      private final Set<E> delegate;
-
-      MutatabilityAwareSet(MutabilityOracle mutabilityOracle, Set<E> delegate) {
-        this.mutabilityOracle = mutabilityOracle;
-        this.delegate = delegate;
-      }
-
-      @Override
-      public int size() {
-        return delegate.size();
-      }
-
-      @Override
-      public boolean isEmpty() {
-        return delegate.isEmpty();
-      }
-
-      @Override
-      public boolean contains(Object o) {
-        return delegate.contains(o);
-      }
-
-      @Override
-      public Iterator<E> iterator() {
-        return new MutatabilityAwareIterator<E>(mutabilityOracle, delegate.iterator());
-      }
-
-      @Override
-      public Object[] toArray() {
-        return delegate.toArray();
-      }
-
-      @Override
-      public <T> T[] toArray(T[] a) {
-        return delegate.toArray(a);
-      }
-
-      @Override
-      public boolean add(E e) {
-        mutabilityOracle.ensureMutable();
-        return delegate.add(e);
-      }
-
-      @Override
-      public boolean remove(Object o) {
-        mutabilityOracle.ensureMutable();
-        return delegate.remove(o);
-      }
-
-      @Override
-      public boolean containsAll(Collection<?> c) {
-        return delegate.containsAll(c);
-      }
-
-      @Override
-      public boolean addAll(Collection<? extends E> c) {
-        mutabilityOracle.ensureMutable();
-        return delegate.addAll(c);
-      }
-
-      @Override
-      public boolean retainAll(Collection<?> c) {
-        mutabilityOracle.ensureMutable();
-        return delegate.retainAll(c);
-      }
-
-      @Override
-      public boolean removeAll(Collection<?> c) {
-        mutabilityOracle.ensureMutable();
-        return delegate.removeAll(c);
-      }
-
-      @Override
-      public void clear() {
-        mutabilityOracle.ensureMutable();
-        delegate.clear();
-      }
-
-      @Override
-      public boolean equals(Object o) {
-        return delegate.equals(o);
-      }
-
-      @Override
-      public int hashCode() {
-        return delegate.hashCode();
-      }
-
-      @Override
-      public String toString() {
-        return delegate.toString();
-      }
-    }
-
-    /**
-     * An internal iterator that checks for mutability before delegating.
-     */
-    private static class MutatabilityAwareIterator<E> implements Iterator<E> {
-      private final MutabilityOracle mutabilityOracle;
-      private final Iterator<E> delegate;
-
-      MutatabilityAwareIterator(MutabilityOracle mutabilityOracle, Iterator<E> delegate) {
-        this.mutabilityOracle = mutabilityOracle;
-        this.delegate = delegate;
-      }
-
-      @Override
-      public boolean hasNext() {
-        return delegate.hasNext();
-      }
-
-      @Override
-      public E next() {
-        return delegate.next();
-      }
-
-      @Override
-      public void remove() {
-        mutabilityOracle.ensureMutable();
-        delegate.remove();
-      }
-
-      @Override
-      public boolean equals(Object obj) {
-        return delegate.equals(obj);
-      }
-
-      @Override
-      public int hashCode() {
-        return delegate.hashCode();
-      }
-
-      @Override
-      public String toString() {
-        return delegate.toString();
-      }
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapFieldLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapFieldLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapFieldLite.java
deleted file mode 100644
index 761c4ce..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MapFieldLite.java
+++ /dev/null
@@ -1,224 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLite;
-
-import java.util.Arrays;
-import java.util.Collections;
-import java.util.LinkedHashMap;
-import java.util.Map;
-import java.util.Set;
-
-/**
- * Internal representation of map fields in generated lite-runtime messages.
- *
- * This class is a protobuf implementation detail. Users shouldn't use this
- * class directly.
- */
-public final class MapFieldLite<K, V> extends LinkedHashMap<K, V> {
-
-  private boolean isMutable;
-
-  private MapFieldLite() {
-    this.isMutable = true;
-  }
-
-  private MapFieldLite(Map<K, V> mapData) {
-    super(mapData);
-    this.isMutable = true;
-  }
-
-  @SuppressWarnings({"rawtypes", "unchecked"})
-  private static final MapFieldLite EMPTY_MAP_FIELD = new MapFieldLite();
-  static {
-    EMPTY_MAP_FIELD.makeImmutable();
-  }
-
-  /** Returns an singleton immutable empty MapFieldLite instance. */
-  @SuppressWarnings({"unchecked", "cast"})
-  public static <K, V> MapFieldLite<K, V> emptyMapField() {
-    return (MapFieldLite<K, V>) EMPTY_MAP_FIELD;
-  }
-
-  public void mergeFrom(MapFieldLite<K, V> other) {
-    ensureMutable();
-    if (!other.isEmpty()) {
-      putAll(other);
-    }
-  }
-
-  @SuppressWarnings({"unchecked", "cast"})
-  @Override public Set<Map.Entry<K, V>> entrySet() {
-    return isEmpty() ? Collections.<Map.Entry<K, V>>emptySet() : super.entrySet();
-  }
-
-  @Override public void clear() {
-    ensureMutable();
-    super.clear();
-  }
-
-  @Override public V put(K key, V value) {
-    ensureMutable();
-    return super.put(key, value);
-  }
-
-  public V put(Map.Entry<K, V> entry) {
-    return put(entry.getKey(), entry.getValue());
-  }
-
-  @Override public void putAll(Map<? extends K, ? extends V> m) {
-    ensureMutable();
-    super.putAll(m);
-  }
-
-  @Override public V remove(Object key) {
-    ensureMutable();
-    return super.remove(key);
-  }
-
-  private static boolean equals(Object a, Object b) {
-    if (a instanceof byte[] && b instanceof byte[]) {
-      return Arrays.equals((byte[]) a, (byte[]) b);
-    }
-    return a.equals(b);
-  }
-
-  /**
-   * Checks whether two {@link Map}s are equal. We don't use the default equals
-   * method of {@link Map} because it compares by identity not by content for
-   * byte arrays.
-   */
-  static <K, V> boolean equals(Map<K, V> a, Map<K, V> b) {
-    if (a == b) {
-      return true;
-    }
-    if (a.size() != b.size()) {
-      return false;
-    }
-    for (Map.Entry<K, V> entry : a.entrySet()) {
-      if (!b.containsKey(entry.getKey())) {
-        return false;
-      }
-      if (!equals(entry.getValue(), b.get(entry.getKey()))) {
-        return false;
-      }
-    }
-    return true;
-  }
-
-  /**
-   * Checks whether two map fields are equal.
-   */
-  @SuppressWarnings("unchecked")
-  @Override
-  public boolean equals(Object object) {
-    return (object instanceof Map) && equals(this, (Map<K, V>) object);
-  }
-
-  private static int calculateHashCodeForObject(Object a) {
-    if (a instanceof byte[]) {
-      return Internal.hashCode((byte[]) a);
-    }
-    // Enums should be stored as integers internally.
-    if (a instanceof EnumLite) {
-      throw new UnsupportedOperationException();
-    }
-    return a.hashCode();
-  }
-
-  /**
-   * Calculates the hash code for a {@link Map}. We don't use the default hash
-   * code method of {@link Map} because for byte arrays and protobuf enums it
-   * use {@link Object#hashCode()}.
-   */
-  static <K, V> int calculateHashCodeForMap(Map<K, V> a) {
-    int result = 0;
-    for (Map.Entry<K, V> entry : a.entrySet()) {
-      result += calculateHashCodeForObject(entry.getKey())
-          ^ calculateHashCodeForObject(entry.getValue());
-    }
-    return result;
-  }
-
-  @Override
-  public int hashCode() {
-    return calculateHashCodeForMap(this);
-  }
-
-  private static Object copy(Object object) {
-    if (object instanceof byte[]) {
-      byte[] data = (byte[]) object;
-      return Arrays.copyOf(data, data.length);
-    }
-    return object;
-  }
-
-  /**
-   * Makes a deep copy of a {@link Map}. Immutable objects in the map will be
-   * shared (e.g., integers, strings, immutable messages) and mutable ones will
-   * have a copy (e.g., byte arrays, mutable messages).
-   */
-  @SuppressWarnings("unchecked")
-  static <K, V> Map<K, V> copy(Map<K, V> map) {
-    Map<K, V> result = new LinkedHashMap<K, V>();
-    for (Map.Entry<K, V> entry : map.entrySet()) {
-      result.put(entry.getKey(), (V) copy(entry.getValue()));
-    }
-    return result;
-  }
-
-  /** Returns a deep copy of this map field. */
-  public MapFieldLite<K, V> mutableCopy() {
-    return isEmpty() ? new MapFieldLite<K, V>() : new MapFieldLite<K, V>(this);
-  }
-
-  /**
-   * Makes this field immutable. All subsequent modifications will throw an
-   * {@link UnsupportedOperationException}.
-   */
-  public void makeImmutable() {
-    isMutable = false;
-  }
-
-  /**
-   * Returns whether this field can be modified.
-   */
-  public boolean isMutable() {
-    return isMutable;
-  }
-
-  private void ensureMutable() {
-    if (!isMutable()) {
-      throw new UnsupportedOperationException();
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Message.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Message.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Message.java
deleted file mode 100644
index e85079d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Message.java
+++ /dev/null
@@ -1,292 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-// TODO(kenton):  Use generics?  E.g. Builder<BuilderType extends Builder>, then
-//   mergeFrom*() could return BuilderType for better type-safety.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.util.Map;
-
-/**
- * Abstract interface implemented by Protocol Message objects.
- * <p>
- * See also {@link MessageLite}, which defines most of the methods that typical
- * users care about.  {@link Message} adds to it methods that are not available
- * in the "lite" runtime.  The biggest added features are introspection and
- * reflection -- i.e., getting descriptors for the message type and accessing
- * the field values dynamically.
- *
- * @author kenton@google.com Kenton Varda
- */
-public interface Message extends MessageLite, MessageOrBuilder {
-
-  // (From MessageLite, re-declared here only for return type covariance.)
-  @Override
-  Parser<? extends Message> getParserForType();
-
-
-  // -----------------------------------------------------------------
-  // Comparison and hashing
-
-  /**
-   * Compares the specified object with this message for equality.  Returns
-   * {@code true} if the given object is a message of the same type (as
-   * defined by {@code getDescriptorForType()}) and has identical values for
-   * all of its fields.  Subclasses must implement this; inheriting
-   * {@code Object.equals()} is incorrect.
-   *
-   * @param other object to be compared for equality with this message
-   * @return {@code true} if the specified object is equal to this message
-   */
-  @Override
-  boolean equals(Object other);
-
-  /**
-   * Returns the hash code value for this message.  The hash code of a message
-   * should mix the message's type (object identity of the descriptor) with its
-   * contents (known and unknown field values).  Subclasses must implement this;
-   * inheriting {@code Object.hashCode()} is incorrect.
-   *
-   * @return the hash code value for this message
-   * @see Map#hashCode()
-   */
-  @Override
-  int hashCode();
-
-  // -----------------------------------------------------------------
-  // Convenience methods.
-
-  /**
-   * Converts the message to a string in protocol buffer text format. This is
-   * just a trivial wrapper around {@link
-   * TextFormat#printToString(MessageOrBuilder)}.
-   */
-  @Override
-  String toString();
-
-  // =================================================================
-  // Builders
-
-  // (From MessageLite, re-declared here only for return type covariance.)
-  @Override
-  Builder newBuilderForType();
-
-  @Override
-  Builder toBuilder();
-
-  /**
-   * Abstract interface implemented by Protocol Message builders.
-   */
-  interface Builder extends MessageLite.Builder, MessageOrBuilder {
-    // (From MessageLite.Builder, re-declared here only for return type
-    // covariance.)
-    @Override
-    Builder clear();
-
-    /**
-     * Merge {@code other} into the message being built.  {@code other} must
-     * have the exact same type as {@code this} (i.e.
-     * {@code getDescriptorForType() == other.getDescriptorForType()}).
-     *
-     * Merging occurs as follows.  For each field:<br>
-     * * For singular primitive fields, if the field is set in {@code other},
-     *   then {@code other}'s value overwrites the value in this message.<br>
-     * * For singular message fields, if the field is set in {@code other},
-     *   it is merged into the corresponding sub-message of this message
-     *   using the same merging rules.<br>
-     * * For repeated fields, the elements in {@code other} are concatenated
-     *   with the elements in this message.
-     * * For oneof groups, if the other message has one of the fields set,
-     *   the group of this message is cleared and replaced by the field
-     *   of the other message, so that the oneof constraint is preserved.
-     *
-     * This is equivalent to the {@code Message::MergeFrom} method in C++.
-     */
-    Builder mergeFrom(Message other);
-
-    // (From MessageLite.Builder, re-declared here only for return type
-    // covariance.)
-    @Override
-    Message build();
-
-    @Override
-    Message buildPartial();
-
-    @Override
-    Builder clone();
-
-    @Override
-    Builder mergeFrom(CodedInputStream input) throws IOException;
-
-    @Override
-    Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-        throws IOException;
-
-    /**
-     * Get the message's type's descriptor.
-     * See {@link Message#getDescriptorForType()}.
-     */
-    @Override
-    Descriptors.Descriptor getDescriptorForType();
-
-    /**
-     * Create a Builder for messages of the appropriate type for the given
-     * field.  Messages built with this can then be passed to setField(),
-     * setRepeatedField(), or addRepeatedField().
-     */
-    Builder newBuilderForField(Descriptors.FieldDescriptor field);
-
-    /**
-     * Get a nested builder instance for the given field.
-     * <p>
-     * Normally, we hold a reference to the immutable message object for the
-     * message type field. Some implementations(the generated message builders),
-     * however, can also hold a reference to the builder object (a nested
-     * builder) for the field.
-     * <p>
-     * If the field is already backed up by a nested builder, the nested builder
-     * will be returned. Otherwise, a new field builder will be created and
-     * returned. The original message field (if exist) will be merged into the
-     * field builder, which will then be nested into its parent builder.
-     * <p>
-     * NOTE: implementations that do not support nested builders will throw
-     * <code>UnsupportedOperationException</code>.
-     */
-    Builder getFieldBuilder(Descriptors.FieldDescriptor field);
-
-    /**
-     * Get a nested builder instance for the given repeated field instance.
-     * <p>
-     * Normally, we hold a reference to the immutable message object for the
-     * message type field. Some implementations(the generated message builders),
-     * however, can also hold a reference to the builder object (a nested
-     * builder) for the field.
-     * <p>
-     * If the field is already backed up by a nested builder, the nested builder
-     * will be returned. Otherwise, a new field builder will be created and
-     * returned. The original message field (if exist) will be merged into the
-     * field builder, which will then be nested into its parent builder.
-     * <p>
-     * NOTE: implementations that do not support nested builders will throw
-     * <code>UnsupportedOperationException</code>.
-     */
-    Builder getRepeatedFieldBuilder(Descriptors.FieldDescriptor field,
-                                    int index);
-
-    /**
-     * Sets a field to the given value.  The value must be of the correct type
-     * for this field, i.e. the same type that
-     * {@link Message#getField(Descriptors.FieldDescriptor)} would return.
-     */
-    Builder setField(Descriptors.FieldDescriptor field, Object value);
-
-    /**
-     * Clears the field.  This is exactly equivalent to calling the generated
-     * "clear" accessor method corresponding to the field.
-     */
-    Builder clearField(Descriptors.FieldDescriptor field);
-
-    /**
-     * Clears the oneof.  This is exactly equivalent to calling the generated
-     * "clear" accessor method corresponding to the oneof.
-     */
-    Builder clearOneof(Descriptors.OneofDescriptor oneof);
-
-    /**
-     * Sets an element of a repeated field to the given value.  The value must
-     * be of the correct type for this field, i.e. the same type that
-     * {@link Message#getRepeatedField(Descriptors.FieldDescriptor,int)} would
-     * return.
-     * @throws IllegalArgumentException The field is not a repeated field, or
-     *           {@code field.getContainingType() != getDescriptorForType()}.
-     */
-    Builder setRepeatedField(Descriptors.FieldDescriptor field,
-                             int index, Object value);
-
-    /**
-     * Like {@code setRepeatedField}, but appends the value as a new element.
-     * @throws IllegalArgumentException The field is not a repeated field, or
-     *           {@code field.getContainingType() != getDescriptorForType()}.
-     */
-    Builder addRepeatedField(Descriptors.FieldDescriptor field, Object value);
-
-    /** Set the {@link UnknownFieldSet} for this message. */
-    Builder setUnknownFields(UnknownFieldSet unknownFields);
-
-    /**
-     * Merge some unknown fields into the {@link UnknownFieldSet} for this
-     * message.
-     */
-    Builder mergeUnknownFields(UnknownFieldSet unknownFields);
-
-    // ---------------------------------------------------------------
-    // Convenience methods.
-
-    // (From MessageLite.Builder, re-declared here only for return type
-    // covariance.)
-    @Override
-    Builder mergeFrom(ByteString data) throws InvalidProtocolBufferException;
-
-    @Override
-    Builder mergeFrom(ByteString data, ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException;
-
-    @Override
-    Builder mergeFrom(byte[] data) throws InvalidProtocolBufferException;
-
-    @Override
-    Builder mergeFrom(byte[] data, int off, int len) throws InvalidProtocolBufferException;
-
-    @Override
-    Builder mergeFrom(byte[] data, ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException;
-
-    @Override
-    Builder mergeFrom(byte[] data, int off, int len, ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException;
-
-    @Override
-    Builder mergeFrom(InputStream input) throws IOException;
-
-    @Override
-    Builder mergeFrom(InputStream input, ExtensionRegistryLite extensionRegistry)
-        throws IOException;
-
-    @Override
-    boolean mergeDelimitedFrom(InputStream input) throws IOException;
-
-    @Override
-    boolean mergeDelimitedFrom(InputStream input, ExtensionRegistryLite extensionRegistry)
-        throws IOException;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLite.java
deleted file mode 100644
index 8340abab..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLite.java
+++ /dev/null
@@ -1,341 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-// TODO(kenton):  Use generics?  E.g. Builder<BuilderType extends Builder>, then
-//   mergeFrom*() could return BuilderType for better type-safety.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.OutputStream;
-
-/**
- * Abstract interface implemented by Protocol Message objects.
- *
- * <p>This interface is implemented by all protocol message objects.  Non-lite
- * messages additionally implement the Message interface, which is a subclass
- * of MessageLite.  Use MessageLite instead when you only need the subset of
- * features which it supports -- namely, nothing that uses descriptors or
- * reflection.  You can instruct the protocol compiler to generate classes
- * which implement only MessageLite, not the full Message interface, by adding
- * the follow line to the .proto file:
- * <pre>
- *   option optimize_for = LITE_RUNTIME;
- * </pre>
- *
- * <p>This is particularly useful on resource-constrained systems where the
- * full protocol buffers runtime library is too big.
- *
- * <p>Note that on non-constrained systems (e.g. servers) when you need to link
- * in lots of protocol definitions, a better way to reduce total code footprint
- * is to use {@code optimize_for = CODE_SIZE}.  This will make the generated
- * code smaller while still supporting all the same features (at the expense of
- * speed).  {@code optimize_for = LITE_RUNTIME} is best when you only have a
- * small number of message types linked into your binary, in which case the
- * size of the protocol buffers runtime itself is the biggest problem.
- *
- * @author kenton@google.com Kenton Varda
- */
-public interface MessageLite extends MessageLiteOrBuilder {
-
-
-  /**
-   * Serializes the message and writes it to {@code output}.  This does not
-   * flush or close the stream.
-   */
-  void writeTo(CodedOutputStream output) throws IOException;
-
-  /**
-   * Get the number of bytes required to encode this message.  The result
-   * is only computed on the first call and memoized after that.
-   */
-  int getSerializedSize();
-
-
-  /**
-   * Gets the parser for a message of the same type as this message.
-   */
-  Parser<? extends MessageLite> getParserForType();
-
-  // -----------------------------------------------------------------
-  // Convenience methods.
-
-  /**
-   * Serializes the message to a {@code ByteString} and returns it. This is
-   * just a trivial wrapper around
-   * {@link #writeTo(CodedOutputStream)}.
-   */
-  ByteString toByteString();
-
-  /**
-   * Serializes the message to a {@code byte} array and returns it.  This is
-   * just a trivial wrapper around
-   * {@link #writeTo(CodedOutputStream)}.
-   */
-  byte[] toByteArray();
-
-  /**
-   * Serializes the message and writes it to {@code output}.  This is just a
-   * trivial wrapper around {@link #writeTo(CodedOutputStream)}.  This does
-   * not flush or close the stream.
-   * <p>
-   * NOTE:  Protocol Buffers are not self-delimiting.  Therefore, if you write
-   * any more data to the stream after the message, you must somehow ensure
-   * that the parser on the receiving end does not interpret this as being
-   * part of the protocol message.  This can be done e.g. by writing the size
-   * of the message before the data, then making sure to limit the input to
-   * that size on the receiving end (e.g. by wrapping the InputStream in one
-   * which limits the input).  Alternatively, just use
-   * {@link #writeDelimitedTo(OutputStream)}.
-   */
-  void writeTo(OutputStream output) throws IOException;
-
-  /**
-   * Like {@link #writeTo(OutputStream)}, but writes the size of the message
-   * as a varint before writing the data.  This allows more data to be written
-   * to the stream after the message without the need to delimit the message
-   * data yourself.  Use {@link Builder#mergeDelimitedFrom(InputStream)} (or
-   * the static method {@code YourMessageType.parseDelimitedFrom(InputStream)})
-   * to parse messages written by this method.
-   */
-  void writeDelimitedTo(OutputStream output) throws IOException;
-
-
-  // =================================================================
-  // Builders
-
-  /**
-   * Constructs a new builder for a message of the same type as this message.
-   */
-  Builder newBuilderForType();
-
-  /**
-   * Constructs a builder initialized with the current message.  Use this to
-   * derive a new message from the current one.
-   */
-  Builder toBuilder();
-
-  /**
-   * Abstract interface implemented by Protocol Message builders.
-   */
-  interface Builder extends MessageLiteOrBuilder, Cloneable {
-    /** Resets all fields to their default values. */
-    Builder clear();
-
-    /**
-     * Constructs the message based on the state of the Builder. Subsequent
-     * changes to the Builder will not affect the returned message.
-     * @throws UninitializedMessageException The message is missing one or more
-     *         required fields (i.e. {@link #isInitialized()} returns false).
-     *         Use {@link #buildPartial()} to bypass this check.
-     */
-    MessageLite build();
-
-    /**
-     * Like {@link #build()}, but does not throw an exception if the message
-     * is missing required fields.  Instead, a partial message is returned.
-     * Subsequent changes to the Builder will not affect the returned message.
-     */
-    MessageLite buildPartial();
-
-    /**
-     * Clones the Builder.
-     * @see Object#clone()
-     */
-    Builder clone();
-
-    /**
-     * Parses a message of this type from the input and merges it with this
-     * message.
-     *
-     * <p>Warning:  This does not verify that all required fields are present in
-     * the input message.  If you call {@link #build()} without setting all
-     * required fields, it will throw an {@link UninitializedMessageException},
-     * which is a {@code RuntimeException} and thus might not be caught.  There
-     * are a few good ways to deal with this:
-     * <ul>
-     *   <li>Call {@link #isInitialized()} to verify that all required fields
-     *       are set before building.
-     *   <li>Use {@code buildPartial()} to build, which ignores missing
-     *       required fields.
-     * </ul>
-     *
-     * <p>Note:  The caller should call
-     * {@link CodedInputStream#checkLastTagWas(int)} after calling this to
-     * verify that the last tag seen was the appropriate end-group tag,
-     * or zero for EOF.
-     */
-    Builder mergeFrom(CodedInputStream input) throws IOException;
-
-    /**
-     * Like {@link Builder#mergeFrom(CodedInputStream)}, but also
-     * parses extensions.  The extensions that you want to be able to parse
-     * must be registered in {@code extensionRegistry}.  Extensions not in
-     * the registry will be treated as unknown fields.
-     */
-    Builder mergeFrom(CodedInputStream input,
-                      ExtensionRegistryLite extensionRegistry)
-                      throws IOException;
-
-    // ---------------------------------------------------------------
-    // Convenience methods.
-
-    /**
-     * Parse {@code data} as a message of this type and merge it with the
-     * message being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream)}.
-     *
-     * @return this
-     */
-    Builder mergeFrom(ByteString data) throws InvalidProtocolBufferException;
-
-    /**
-     * Parse {@code data} as a message of this type and merge it with the
-     * message being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream,ExtensionRegistryLite)}.
-     *
-     * @return this
-     */
-    Builder mergeFrom(ByteString data,
-                      ExtensionRegistryLite extensionRegistry)
-                      throws InvalidProtocolBufferException;
-
-    /**
-     * Parse {@code data} as a message of this type and merge it with the
-     * message being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream)}.
-     *
-     * @return this
-     */
-    Builder mergeFrom(byte[] data) throws InvalidProtocolBufferException;
-
-    /**
-     * Parse {@code data} as a message of this type and merge it with the
-     * message being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream)}.
-     *
-     * @return this
-     */
-    Builder mergeFrom(byte[] data, int off, int len)
-                      throws InvalidProtocolBufferException;
-
-    /**
-     * Parse {@code data} as a message of this type and merge it with the
-     * message being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream,ExtensionRegistryLite)}.
-     *
-     * @return this
-     */
-    Builder mergeFrom(byte[] data,
-                      ExtensionRegistryLite extensionRegistry)
-                      throws InvalidProtocolBufferException;
-
-    /**
-     * Parse {@code data} as a message of this type and merge it with the
-     * message being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream,ExtensionRegistryLite)}.
-     *
-     * @return this
-     */
-    Builder mergeFrom(byte[] data, int off, int len,
-                      ExtensionRegistryLite extensionRegistry)
-                      throws InvalidProtocolBufferException;
-
-    /**
-     * Parse a message of this type from {@code input} and merge it with the
-     * message being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream)}.  Note that this method always
-     * reads the <i>entire</i> input (unless it throws an exception).  If you
-     * want it to stop earlier, you will need to wrap your input in some
-     * wrapper stream that limits reading.  Or, use
-     * {@link MessageLite#writeDelimitedTo(OutputStream)} to write your message
-     * and {@link #mergeDelimitedFrom(InputStream)} to read it.
-     * <p>
-     * Despite usually reading the entire input, this does not close the stream.
-     *
-     * @return this
-     */
-    Builder mergeFrom(InputStream input) throws IOException;
-
-    /**
-     * Parse a message of this type from {@code input} and merge it with the
-     * message being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream,ExtensionRegistryLite)}.
-     *
-     * @return this
-     */
-    Builder mergeFrom(InputStream input,
-                      ExtensionRegistryLite extensionRegistry)
-                      throws IOException;
-    
-    /**
-     * Merge {@code other} into the message being built.  {@code other} must
-     * have the exact same type as {@code this} (i.e.
-     * {@code getClass().equals(getDefaultInstanceForType().getClass())}).
-     *
-     * Merging occurs as follows.  For each field:<br>
-     * * For singular primitive fields, if the field is set in {@code other},
-     *   then {@code other}'s value overwrites the value in this message.<br>
-     * * For singular message fields, if the field is set in {@code other},
-     *   it is merged into the corresponding sub-message of this message
-     *   using the same merging rules.<br>
-     * * For repeated fields, the elements in {@code other} are concatenated
-     *   with the elements in this message.
-     * * For oneof groups, if the other message has one of the fields set,
-     *   the group of this message is cleared and replaced by the field
-     *   of the other message, so that the oneof constraint is preserved.
-     *
-     * This is equivalent to the {@code Message::MergeFrom} method in C++.
-     */
-    Builder mergeFrom(MessageLite other);
-
-    /**
-     * Like {@link #mergeFrom(InputStream)}, but does not read until EOF.
-     * Instead, the size of the message (encoded as a varint) is read first,
-     * then the message data.  Use
-     * {@link MessageLite#writeDelimitedTo(OutputStream)} to write messages in
-     * this format.
-     *
-     * @return True if successful, or false if the stream is at EOF when the
-     *         method starts.  Any other error (including reaching EOF during
-     *         parsing) will cause an exception to be thrown.
-     */
-    boolean mergeDelimitedFrom(InputStream input)
-                               throws IOException;
-
-    /**
-     * Like {@link #mergeDelimitedFrom(InputStream)} but supporting extensions.
-     */
-    boolean mergeDelimitedFrom(InputStream input,
-                               ExtensionRegistryLite extensionRegistry)
-                               throws IOException;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLiteOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLiteOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLiteOrBuilder.java
deleted file mode 100644
index 39f144d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLiteOrBuilder.java
+++ /dev/null
@@ -1,60 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Base interface for methods common to {@link MessageLite}
- * and {@link MessageLite.Builder} to provide type equivalency.
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public interface MessageLiteOrBuilder {
-  /**
-   * Get an instance of the type with no fields set. Because no fields are set,
-   * all getters for singular fields will return default values and repeated
-   * fields will appear empty.
-   * This may or may not be a singleton.  This differs from the
-   * {@code getDefaultInstance()} method of generated message classes in that
-   * this method is an abstract method of the {@code MessageLite} interface
-   * whereas {@code getDefaultInstance()} is a static method of a specific
-   * class.  They return the same thing.
-   */
-  MessageLite getDefaultInstanceForType();
-
-  /**
-   * Returns true if all required fields in the message and all embedded
-   * messages are set, false otherwise.
-   *
-   * <p>See also: {@link MessageOrBuilder#getInitializationErrorString()}
-   */
-  boolean isInitialized();
-
-}


[32/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleArrayList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleArrayList.java
deleted file mode 100644
index 88effb2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleArrayList.java
+++ /dev/null
@@ -1,273 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.DoubleList;
-
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.RandomAccess;
-
-/**
- * An implementation of {@link DoubleList} on top of a primitive array.
- *
- * @author dweis@google.com (Daniel Weis)
- */
-final class DoubleArrayList
-    extends AbstractProtobufList<Double>
-    implements DoubleList, RandomAccess {
-
-  private static final DoubleArrayList EMPTY_LIST = new DoubleArrayList();
-  static {
-    EMPTY_LIST.makeImmutable();
-  }
-
-  public static DoubleArrayList emptyList() {
-    return EMPTY_LIST;
-  }
-
-  /**
-   * The backing store for the list.
-   */
-  private double[] array;
-
-  /**
-   * The size of the list distinct from the length of the array. That is, it is the number of
-   * elements set in the list.
-   */
-  private int size;
-
-  /**
-   * Constructs a new mutable {@code DoubleArrayList} with default capacity.
-   */
-  DoubleArrayList() {
-    this(new double[DEFAULT_CAPACITY], 0);
-  }
-
-  /**
-   * Constructs a new mutable {@code DoubleArrayList}
-   * containing the same elements as {@code other}.
-   */
-  private DoubleArrayList(double[] other, int size) {
-    array = other;
-    this.size = size;
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (this == o) {
-      return true;
-    }
-    if (!(o instanceof DoubleArrayList)) {
-      return super.equals(o);
-    }
-    DoubleArrayList other = (DoubleArrayList) o;
-    if (size != other.size) {
-      return false;
-    }
-
-    final double[] arr = other.array;
-    for (int i = 0; i < size; i++) {
-      if (array[i] != arr[i]) {
-        return false;
-      }
-    }
-
-    return true;
-  }
-
-  @Override
-  public int hashCode() {
-    int result = 1;
-    for (int i = 0; i < size; i++) {
-      long bits = Double.doubleToLongBits(array[i]);
-      result = (31 * result) + Internal.hashLong(bits);
-    }
-    return result;
-  }
-
-  @Override
-  public DoubleList mutableCopyWithCapacity(int capacity) {
-    if (capacity < size) {
-      throw new IllegalArgumentException();
-    }
-    return new DoubleArrayList(Arrays.copyOf(array, capacity), size);
-  }
-
-  @Override
-  public Double get(int index) {
-    return getDouble(index);
-  }
-
-  @Override
-  public double getDouble(int index) {
-    ensureIndexInRange(index);
-    return array[index];
-  }
-
-  @Override
-  public int size() {
-    return size;
-  }
-
-  @Override
-  public Double set(int index, Double element) {
-    return setDouble(index, element);
-  }
-
-  @Override
-  public double setDouble(int index, double element) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    double previousValue = array[index];
-    array[index] = element;
-    return previousValue;
-  }
-
-  @Override
-  public void add(int index, Double element) {
-    addDouble(index, element);
-  }
-
-  /**
-   * Like {@link #add(Double)} but more efficient in that it doesn't box the element.
-   */
-  @Override
-  public void addDouble(double element) {
-    addDouble(size, element);
-  }
-
-  /**
-   * Like {@link #add(int, Double)} but more efficient in that it doesn't box the element.
-   */
-  private void addDouble(int index, double element) {
-    ensureIsMutable();
-    if (index < 0 || index > size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-
-    if (size < array.length) {
-      // Shift everything over to make room
-      System.arraycopy(array, index, array, index + 1, size - index);
-    } else {
-      // Resize to 1.5x the size
-      int length = ((size * 3) / 2) + 1;
-      double[] newArray = new double[length];
-
-      // Copy the first part directly
-      System.arraycopy(array, 0, newArray, 0, index);
-
-      // Copy the rest shifted over by one to make room
-      System.arraycopy(array, index, newArray, index + 1, size - index);
-      array = newArray;
-    }
-
-    array[index] = element;
-    size++;
-    modCount++;
-  }
-
-  @Override
-  public boolean addAll(Collection<? extends Double> collection) {
-    ensureIsMutable();
-
-    if (collection == null) {
-      throw new NullPointerException();
-    }
-
-    // We specialize when adding another DoubleArrayList to avoid boxing elements.
-    if (!(collection instanceof DoubleArrayList)) {
-      return super.addAll(collection);
-    }
-
-    DoubleArrayList list = (DoubleArrayList) collection;
-    if (list.size == 0) {
-      return false;
-    }
-
-    int overflow = Integer.MAX_VALUE - size;
-    if (overflow < list.size) {
-      // We can't actually represent a list this large.
-      throw new OutOfMemoryError();
-    }
-
-    int newSize = size + list.size;
-    if (newSize > array.length) {
-      array = Arrays.copyOf(array, newSize);
-    }
-
-    System.arraycopy(list.array, 0, array, size, list.size);
-    size = newSize;
-    modCount++;
-    return true;
-  }
-
-  @Override
-  public boolean remove(Object o) {
-    ensureIsMutable();
-    for (int i = 0; i < size; i++) {
-      if (o.equals(array[i])) {
-        System.arraycopy(array, i + 1, array, i, size - i);
-        size--;
-        modCount++;
-        return true;
-      }
-    }
-    return false;
-  }
-
-  @Override
-  public Double remove(int index) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    double value = array[index];
-    System.arraycopy(array, index + 1, array, index, size - index);
-    size--;
-    modCount++;
-    return value;
-  }
-
-  /**
-   * Ensures that the provided {@code index} is within the range of {@code [0, size]}. Throws an
-   * {@link IndexOutOfBoundsException} if it is not.
-   *
-   * @param index the index to verify is in range
-   */
-  private void ensureIndexInRange(int index) {
-    if (index < 0 || index >= size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-  }
-
-  private String makeOutOfBoundsExceptionMessage(int index) {
-    return "Index:" + index + ", Size:" + size;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleValue.java
deleted file mode 100644
index 98a0ecc..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleValue.java
+++ /dev/null
@@ -1,454 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `double`.
- * The JSON representation for `DoubleValue` is JSON number.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.DoubleValue}
- */
-public  final class DoubleValue extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.DoubleValue)
-    DoubleValueOrBuilder {
-  // Use DoubleValue.newBuilder() to construct.
-  private DoubleValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private DoubleValue() {
-    value_ = 0D;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private DoubleValue(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 9: {
-
-            value_ = input.readDouble();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_DoubleValue_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_DoubleValue_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private double value_;
-  /**
-   * <pre>
-   * The double value.
-   * </pre>
-   *
-   * <code>double value = 1;</code>
-   */
-  public double getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (value_ != 0D) {
-      output.writeDouble(1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (value_ != 0D) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeDoubleSize(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue) obj;
-
-    boolean result = true;
-    result = result && (
-        java.lang.Double.doubleToLongBits(getValue())
-        == java.lang.Double.doubleToLongBits(
-            other.getValue()));
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
-        java.lang.Double.doubleToLongBits(getValue()));
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `double`.
-   * The JSON representation for `DoubleValue` is JSON number.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.DoubleValue}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.DoubleValue)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_DoubleValue_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_DoubleValue_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = 0D;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_DoubleValue_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue.getDefaultInstance()) return this;
-      if (other.getValue() != 0D) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private double value_ ;
-    /**
-     * <pre>
-     * The double value.
-     * </pre>
-     *
-     * <code>double value = 1;</code>
-     */
-    public double getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * The double value.
-     * </pre>
-     *
-     * <code>double value = 1;</code>
-     */
-    public Builder setValue(double value) {
-      
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The double value.
-     * </pre>
-     *
-     * <code>double value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = 0D;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.DoubleValue)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.DoubleValue)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DoubleValue>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<DoubleValue>() {
-    public DoubleValue parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new DoubleValue(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DoubleValue> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<DoubleValue> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.DoubleValue getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleValueOrBuilder.java
deleted file mode 100644
index 180fa50..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DoubleValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface DoubleValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.DoubleValue)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The double value.
-   * </pre>
-   *
-   * <code>double value = 1;</code>
-   */
-  double getValue();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Duration.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Duration.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Duration.java
deleted file mode 100644
index c6736ef..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Duration.java
+++ /dev/null
@@ -1,618 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/duration.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * A Duration represents a signed, fixed-length span of time represented
- * as a count of seconds and fractions of seconds at nanosecond
- * resolution. It is independent of any calendar and concepts like "day"
- * or "month". It is related to Timestamp in that the difference between
- * two Timestamp values is a Duration and it can be added or subtracted
- * from a Timestamp. Range is approximately +-10,000 years.
- * Example 1: Compute Duration from two Timestamps in pseudo code.
- *     Timestamp start = ...;
- *     Timestamp end = ...;
- *     Duration duration = ...;
- *     duration.seconds = end.seconds - start.seconds;
- *     duration.nanos = end.nanos - start.nanos;
- *     if (duration.seconds &lt; 0 &amp;&amp; duration.nanos &gt; 0) {
- *       duration.seconds += 1;
- *       duration.nanos -= 1000000000;
- *     } else if (durations.seconds &gt; 0 &amp;&amp; duration.nanos &lt; 0) {
- *       duration.seconds -= 1;
- *       duration.nanos += 1000000000;
- *     }
- * Example 2: Compute Timestamp from Timestamp + Duration in pseudo code.
- *     Timestamp start = ...;
- *     Duration duration = ...;
- *     Timestamp end = ...;
- *     end.seconds = start.seconds + duration.seconds;
- *     end.nanos = start.nanos + duration.nanos;
- *     if (end.nanos &lt; 0) {
- *       end.seconds -= 1;
- *       end.nanos += 1000000000;
- *     } else if (end.nanos &gt;= 1000000000) {
- *       end.seconds += 1;
- *       end.nanos -= 1000000000;
- *     }
- * Example 3: Compute Duration from datetime.timedelta in Python.
- *     td = datetime.timedelta(days=3, minutes=10)
- *     duration = Duration()
- *     duration.FromTimedelta(td)
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Duration}
- */
-public  final class Duration extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Duration)
-    DurationOrBuilder {
-  // Use Duration.newBuilder() to construct.
-  private Duration(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Duration() {
-    seconds_ = 0L;
-    nanos_ = 0;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Duration(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-
-            seconds_ = input.readInt64();
-            break;
-          }
-          case 16: {
-
-            nanos_ = input.readInt32();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.DurationProto.internal_static_google_protobuf_Duration_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.DurationProto.internal_static_google_protobuf_Duration_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration.Builder.class);
-  }
-
-  public static final int SECONDS_FIELD_NUMBER = 1;
-  private long seconds_;
-  /**
-   * <pre>
-   * Signed seconds of the span of time. Must be from -315,576,000,000
-   * to +315,576,000,000 inclusive.
-   * </pre>
-   *
-   * <code>int64 seconds = 1;</code>
-   */
-  public long getSeconds() {
-    return seconds_;
-  }
-
-  public static final int NANOS_FIELD_NUMBER = 2;
-  private int nanos_;
-  /**
-   * <pre>
-   * Signed fractions of a second at nanosecond resolution of the span
-   * of time. Durations less than one second are represented with a 0
-   * `seconds` field and a positive or negative `nanos` field. For durations
-   * of one second or more, a non-zero value for the `nanos` field must be
-   * of the same sign as the `seconds` field. Must be from -999,999,999
-   * to +999,999,999 inclusive.
-   * </pre>
-   *
-   * <code>int32 nanos = 2;</code>
-   */
-  public int getNanos() {
-    return nanos_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (seconds_ != 0L) {
-      output.writeInt64(1, seconds_);
-    }
-    if (nanos_ != 0) {
-      output.writeInt32(2, nanos_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (seconds_ != 0L) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt64Size(1, seconds_);
-    }
-    if (nanos_ != 0) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt32Size(2, nanos_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration) obj;
-
-    boolean result = true;
-    result = result && (getSeconds()
-        == other.getSeconds());
-    result = result && (getNanos()
-        == other.getNanos());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + SECONDS_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
-        getSeconds());
-    hash = (37 * hash) + NANOS_FIELD_NUMBER;
-    hash = (53 * hash) + getNanos();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * A Duration represents a signed, fixed-length span of time represented
-   * as a count of seconds and fractions of seconds at nanosecond
-   * resolution. It is independent of any calendar and concepts like "day"
-   * or "month". It is related to Timestamp in that the difference between
-   * two Timestamp values is a Duration and it can be added or subtracted
-   * from a Timestamp. Range is approximately +-10,000 years.
-   * Example 1: Compute Duration from two Timestamps in pseudo code.
-   *     Timestamp start = ...;
-   *     Timestamp end = ...;
-   *     Duration duration = ...;
-   *     duration.seconds = end.seconds - start.seconds;
-   *     duration.nanos = end.nanos - start.nanos;
-   *     if (duration.seconds &lt; 0 &amp;&amp; duration.nanos &gt; 0) {
-   *       duration.seconds += 1;
-   *       duration.nanos -= 1000000000;
-   *     } else if (durations.seconds &gt; 0 &amp;&amp; duration.nanos &lt; 0) {
-   *       duration.seconds -= 1;
-   *       duration.nanos += 1000000000;
-   *     }
-   * Example 2: Compute Timestamp from Timestamp + Duration in pseudo code.
-   *     Timestamp start = ...;
-   *     Duration duration = ...;
-   *     Timestamp end = ...;
-   *     end.seconds = start.seconds + duration.seconds;
-   *     end.nanos = start.nanos + duration.nanos;
-   *     if (end.nanos &lt; 0) {
-   *       end.seconds -= 1;
-   *       end.nanos += 1000000000;
-   *     } else if (end.nanos &gt;= 1000000000) {
-   *       end.seconds += 1;
-   *       end.nanos -= 1000000000;
-   *     }
-   * Example 3: Compute Duration from datetime.timedelta in Python.
-   *     td = datetime.timedelta(days=3, minutes=10)
-   *     duration = Duration()
-   *     duration.FromTimedelta(td)
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Duration}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Duration)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.DurationOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.DurationProto.internal_static_google_protobuf_Duration_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.DurationProto.internal_static_google_protobuf_Duration_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      seconds_ = 0L;
-
-      nanos_ = 0;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.DurationProto.internal_static_google_protobuf_Duration_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration(this);
-      result.seconds_ = seconds_;
-      result.nanos_ = nanos_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration.getDefaultInstance()) return this;
-      if (other.getSeconds() != 0L) {
-        setSeconds(other.getSeconds());
-      }
-      if (other.getNanos() != 0) {
-        setNanos(other.getNanos());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private long seconds_ ;
-    /**
-     * <pre>
-     * Signed seconds of the span of time. Must be from -315,576,000,000
-     * to +315,576,000,000 inclusive.
-     * </pre>
-     *
-     * <code>int64 seconds = 1;</code>
-     */
-    public long getSeconds() {
-      return seconds_;
-    }
-    /**
-     * <pre>
-     * Signed seconds of the span of time. Must be from -315,576,000,000
-     * to +315,576,000,000 inclusive.
-     * </pre>
-     *
-     * <code>int64 seconds = 1;</code>
-     */
-    public Builder setSeconds(long value) {
-      
-      seconds_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Signed seconds of the span of time. Must be from -315,576,000,000
-     * to +315,576,000,000 inclusive.
-     * </pre>
-     *
-     * <code>int64 seconds = 1;</code>
-     */
-    public Builder clearSeconds() {
-      
-      seconds_ = 0L;
-      onChanged();
-      return this;
-    }
-
-    private int nanos_ ;
-    /**
-     * <pre>
-     * Signed fractions of a second at nanosecond resolution of the span
-     * of time. Durations less than one second are represented with a 0
-     * `seconds` field and a positive or negative `nanos` field. For durations
-     * of one second or more, a non-zero value for the `nanos` field must be
-     * of the same sign as the `seconds` field. Must be from -999,999,999
-     * to +999,999,999 inclusive.
-     * </pre>
-     *
-     * <code>int32 nanos = 2;</code>
-     */
-    public int getNanos() {
-      return nanos_;
-    }
-    /**
-     * <pre>
-     * Signed fractions of a second at nanosecond resolution of the span
-     * of time. Durations less than one second are represented with a 0
-     * `seconds` field and a positive or negative `nanos` field. For durations
-     * of one second or more, a non-zero value for the `nanos` field must be
-     * of the same sign as the `seconds` field. Must be from -999,999,999
-     * to +999,999,999 inclusive.
-     * </pre>
-     *
-     * <code>int32 nanos = 2;</code>
-     */
-    public Builder setNanos(int value) {
-      
-      nanos_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Signed fractions of a second at nanosecond resolution of the span
-     * of time. Durations less than one second are represented with a 0
-     * `seconds` field and a positive or negative `nanos` field. For durations
-     * of one second or more, a non-zero value for the `nanos` field must be
-     * of the same sign as the `seconds` field. Must be from -999,999,999
-     * to +999,999,999 inclusive.
-     * </pre>
-     *
-     * <code>int32 nanos = 2;</code>
-     */
-    public Builder clearNanos() {
-      
-      nanos_ = 0;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Duration)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Duration)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Duration>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Duration>() {
-    public Duration parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Duration(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Duration> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Duration> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DurationOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DurationOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DurationOrBuilder.java
deleted file mode 100644
index 9c19aae..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DurationOrBuilder.java
+++ /dev/null
@@ -1,33 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/duration.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface DurationOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Duration)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * Signed seconds of the span of time. Must be from -315,576,000,000
-   * to +315,576,000,000 inclusive.
-   * </pre>
-   *
-   * <code>int64 seconds = 1;</code>
-   */
-  long getSeconds();
-
-  /**
-   * <pre>
-   * Signed fractions of a second at nanosecond resolution of the span
-   * of time. Durations less than one second are represented with a 0
-   * `seconds` field and a positive or negative `nanos` field. For durations
-   * of one second or more, a non-zero value for the `nanos` field must be
-   * of the same sign as the `seconds` field. Must be from -999,999,999
-   * to +999,999,999 inclusive.
-   * </pre>
-   *
-   * <code>int32 nanos = 2;</code>
-   */
-  int getNanos();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DurationProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DurationProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DurationProto.java
deleted file mode 100644
index ee379ea..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DurationProto.java
+++ /dev/null
@@ -1,59 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/duration.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class DurationProto {
-  private DurationProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Duration_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Duration_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\036google/protobuf/duration.proto\022\017google" +
-      ".protobuf\"*\n\010Duration\022\017\n\007seconds\030\001 \001(\003\022\r" +
-      "\n\005nanos\030\002 \001(\005B|\n\023com.google.protobufB\rDu" +
-      "rationProtoP\001Z*github.com/golang/protobu" +
-      "f/ptypes/duration\370\001\001\242\002\003GPB\252\002\036Google.Prot" +
-      "obuf.WellKnownTypesb\006proto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_google_protobuf_Duration_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_Duration_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Duration_descriptor,
-        new java.lang.String[] { "Seconds", "Nanos", });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DynamicMessage.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DynamicMessage.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DynamicMessage.java
deleted file mode 100644
index 13bccd6..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DynamicMessage.java
+++ /dev/null
@@ -1,684 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.util.Collections;
-import java.util.List;
-import java.util.Map;
-
-/**
- * An implementation of {@link Message} that can represent arbitrary types,
- * given a {@link Descriptors.Descriptor}.
- *
- * @author kenton@google.com Kenton Varda
- */
-public final class DynamicMessage extends AbstractMessage {
-  private final Descriptor type;
-  private final FieldSet<FieldDescriptor> fields;
-  private final FieldDescriptor[] oneofCases;
-  private final UnknownFieldSet unknownFields;
-  private int memoizedSize = -1;
-
-  /**
-   * Construct a {@code DynamicMessage} using the given {@code FieldSet}.
-   * oneofCases stores the FieldDescriptor for each oneof to indicate
-   * which field is set. Caller should make sure the array is immutable.
-   *
-   * This constructor is package private and will be used in
-   * {@code DynamicMutableMessage} to convert a mutable message to an immutable
-   * message.
-   */
-  DynamicMessage(Descriptor type, FieldSet<FieldDescriptor> fields,
-                 FieldDescriptor[] oneofCases,
-                 UnknownFieldSet unknownFields) {
-    this.type = type;
-    this.fields = fields;
-    this.oneofCases = oneofCases;
-    this.unknownFields = unknownFields;
-  }
-
-  /**
-   * Get a {@code DynamicMessage} representing the default instance of the
-   * given type.
-   */
-  public static DynamicMessage getDefaultInstance(Descriptor type) {
-    int oneofDeclCount = type.toProto().getOneofDeclCount();
-    FieldDescriptor[] oneofCases = new FieldDescriptor[oneofDeclCount];
-    return new DynamicMessage(type, FieldSet.<FieldDescriptor>emptySet(),
-                              oneofCases,
-                              UnknownFieldSet.getDefaultInstance());
-  }
-
-
-  /** Parse a message of the given type from the given input stream. */
-  public static DynamicMessage parseFrom(Descriptor type,
-                                         CodedInputStream input)
-                                         throws IOException {
-    return newBuilder(type).mergeFrom(input).buildParsed();
-  }
-
-  /** Parse a message of the given type from the given input stream. */
-  public static DynamicMessage parseFrom(
-      Descriptor type,
-      CodedInputStream input,
-      ExtensionRegistry extensionRegistry)
-      throws IOException {
-    return newBuilder(type).mergeFrom(input, extensionRegistry).buildParsed();
-  }
-
-  /** Parse {@code data} as a message of the given type and return it. */
-  public static DynamicMessage parseFrom(Descriptor type, ByteString data)
-                                         throws InvalidProtocolBufferException {
-    return newBuilder(type).mergeFrom(data).buildParsed();
-  }
-
-  /** Parse {@code data} as a message of the given type and return it. */
-  public static DynamicMessage parseFrom(Descriptor type, ByteString data,
-                                         ExtensionRegistry extensionRegistry)
-                                         throws InvalidProtocolBufferException {
-    return newBuilder(type).mergeFrom(data, extensionRegistry).buildParsed();
-  }
-
-  /** Parse {@code data} as a message of the given type and return it. */
-  public static DynamicMessage parseFrom(Descriptor type, byte[] data)
-                                         throws InvalidProtocolBufferException {
-    return newBuilder(type).mergeFrom(data).buildParsed();
-  }
-
-  /** Parse {@code data} as a message of the given type and return it. */
-  public static DynamicMessage parseFrom(Descriptor type, byte[] data,
-                                         ExtensionRegistry extensionRegistry)
-                                         throws InvalidProtocolBufferException {
-    return newBuilder(type).mergeFrom(data, extensionRegistry).buildParsed();
-  }
-
-  /** Parse a message of the given type from {@code input} and return it. */
-  public static DynamicMessage parseFrom(Descriptor type, InputStream input)
-                                         throws IOException {
-    return newBuilder(type).mergeFrom(input).buildParsed();
-  }
-
-  /** Parse a message of the given type from {@code input} and return it. */
-  public static DynamicMessage parseFrom(Descriptor type, InputStream input,
-                                         ExtensionRegistry extensionRegistry)
-                                         throws IOException {
-    return newBuilder(type).mergeFrom(input, extensionRegistry).buildParsed();
-  }
-
-  /** Construct a {@link Message.Builder} for the given type. */
-  public static Builder newBuilder(Descriptor type) {
-    return new Builder(type);
-  }
-
-  /**
-   * Construct a {@link Message.Builder} for a message of the same type as
-   * {@code prototype}, and initialize it with {@code prototype}'s contents.
-   */
-  public static Builder newBuilder(Message prototype) {
-    return new Builder(prototype.getDescriptorForType()).mergeFrom(prototype);
-  }
-
-  // -----------------------------------------------------------------
-  // Implementation of Message interface.
-
-  @Override
-  public Descriptor getDescriptorForType() {
-    return type;
-  }
-
-  @Override
-  public DynamicMessage getDefaultInstanceForType() {
-    return getDefaultInstance(type);
-  }
-
-  @Override
-  public Map<FieldDescriptor, Object> getAllFields() {
-    return fields.getAllFields();
-  }
-
-  @Override
-  public boolean hasOneof(OneofDescriptor oneof) {
-    verifyOneofContainingType(oneof);
-    FieldDescriptor field = oneofCases[oneof.getIndex()];
-    if (field == null) {
-      return false;
-    }
-    return true;
-  }
-
-  @Override
-  public FieldDescriptor getOneofFieldDescriptor(OneofDescriptor oneof) {
-    verifyOneofContainingType(oneof);
-    return oneofCases[oneof.getIndex()];
-  }
-
-  @Override
-  public boolean hasField(FieldDescriptor field) {
-    verifyContainingType(field);
-    return fields.hasField(field);
-  }
-
-  @Override
-  public Object getField(FieldDescriptor field) {
-    verifyContainingType(field);
-    Object result = fields.getField(field);
-    if (result == null) {
-      if (field.isRepeated()) {
-        result = Collections.emptyList();
-      } else if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-        result = getDefaultInstance(field.getMessageType());
-      } else {
-        result = field.getDefaultValue();
-      }
-    }
-    return result;
-  }
-
-  @Override
-  public int getRepeatedFieldCount(FieldDescriptor field) {
-    verifyContainingType(field);
-    return fields.getRepeatedFieldCount(field);
-  }
-
-  @Override
-  public Object getRepeatedField(FieldDescriptor field, int index) {
-    verifyContainingType(field);
-    return fields.getRepeatedField(field, index);
-  }
-
-  @Override
-  public UnknownFieldSet getUnknownFields() {
-    return unknownFields;
-  }
-
-  static boolean isInitialized(Descriptor type,
-                               FieldSet<FieldDescriptor> fields) {
-    // Check that all required fields are present.
-    for (final FieldDescriptor field : type.getFields()) {
-      if (field.isRequired()) {
-        if (!fields.hasField(field)) {
-          return false;
-        }
-      }
-    }
-
-    // Check that embedded messages are initialized.
-    return fields.isInitialized();
-  }
-
-  @Override
-  public boolean isInitialized() {
-    return isInitialized(type, fields);
-  }
-
-  @Override
-  public void writeTo(CodedOutputStream output) throws IOException {
-    if (type.getOptions().getMessageSetWireFormat()) {
-      fields.writeMessageSetTo(output);
-      unknownFields.writeAsMessageSetTo(output);
-    } else {
-      fields.writeTo(output);
-      unknownFields.writeTo(output);
-    }
-  }
-
-  @Override
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    if (type.getOptions().getMessageSetWireFormat()) {
-      size = fields.getMessageSetSerializedSize();
-      size += unknownFields.getSerializedSizeAsMessageSet();
-    } else {
-      size = fields.getSerializedSize();
-      size += unknownFields.getSerializedSize();
-    }
-
-    memoizedSize = size;
-    return size;
-  }
-
-  @Override
-  public Builder newBuilderForType() {
-    return new Builder(type);
-  }
-
-  @Override
-  public Builder toBuilder() {
-    return newBuilderForType().mergeFrom(this);
-  }
-
-  @Override
-  public Parser<DynamicMessage> getParserForType() {
-    return new AbstractParser<DynamicMessage>() {
-      @Override
-      public DynamicMessage parsePartialFrom(
-          CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-          throws InvalidProtocolBufferException {
-        Builder builder = newBuilder(type);
-        try {
-          builder.mergeFrom(input, extensionRegistry);
-        } catch (InvalidProtocolBufferException e) {
-          throw e.setUnfinishedMessage(builder.buildPartial());
-        } catch (IOException e) {
-          throw new InvalidProtocolBufferException(e)
-              .setUnfinishedMessage(builder.buildPartial());
-        }
-        return builder.buildPartial();
-      }
-    };
-  }
-
-  /** Verifies that the field is a field of this message. */
-  private void verifyContainingType(FieldDescriptor field) {
-    if (field.getContainingType() != type) {
-      throw new IllegalArgumentException(
-        "FieldDescriptor does not match message type.");
-    }
-  }
-
-  /** Verifies that the oneof is an oneof of this message. */
-  private void verifyOneofContainingType(OneofDescriptor oneof) {
-    if (oneof.getContainingType() != type) {
-      throw new IllegalArgumentException(
-        "OneofDescriptor does not match message type.");
-    }
-  }
-
-  // =================================================================
-
-  /**
-   * Builder for {@link DynamicMessage}s.
-   */
-  public static final class Builder extends AbstractMessage.Builder<Builder> {
-    private final Descriptor type;
-    private FieldSet<FieldDescriptor> fields;
-    private final FieldDescriptor[] oneofCases;
-    private UnknownFieldSet unknownFields;
-
-    /** Construct a {@code Builder} for the given type. */
-    private Builder(Descriptor type) {
-      this.type = type;
-      this.fields = FieldSet.newFieldSet();
-      this.unknownFields = UnknownFieldSet.getDefaultInstance();
-      this.oneofCases = new FieldDescriptor[type.toProto().getOneofDeclCount()];
-    }
-
-    // ---------------------------------------------------------------
-    // Implementation of Message.Builder interface.
-
-    @Override
-    public Builder clear() {
-      if (fields.isImmutable()) {
-        fields = FieldSet.newFieldSet();
-      } else {
-        fields.clear();
-      }
-      unknownFields = UnknownFieldSet.getDefaultInstance();
-      return this;
-    }
-
-    @Override
-    public Builder mergeFrom(Message other) {
-      if (other instanceof DynamicMessage) {
-        // This should be somewhat faster than calling super.mergeFrom().
-        DynamicMessage otherDynamicMessage = (DynamicMessage) other;
-        if (otherDynamicMessage.type != type) {
-          throw new IllegalArgumentException(
-            "mergeFrom(Message) can only merge messages of the same type.");
-        }
-        ensureIsMutable();
-        fields.mergeFrom(otherDynamicMessage.fields);
-        mergeUnknownFields(otherDynamicMessage.unknownFields);
-        for (int i = 0; i < oneofCases.length; i++) {
-          if (oneofCases[i] == null) {
-            oneofCases[i] = otherDynamicMessage.oneofCases[i];
-          } else {
-            if ((otherDynamicMessage.oneofCases[i] != null)
-                && (oneofCases[i] != otherDynamicMessage.oneofCases[i])) {
-              fields.clearField(oneofCases[i]);
-              oneofCases[i] = otherDynamicMessage.oneofCases[i];
-            }
-          }
-        }
-        return this;
-      } else {
-        return super.mergeFrom(other);
-      }
-    }
-
-    @Override
-    public DynamicMessage build() {
-      if (!isInitialized()) {
-        throw newUninitializedMessageException(
-          new DynamicMessage(type, fields,
-              java.util.Arrays.copyOf(oneofCases, oneofCases.length), unknownFields));
-      }
-      return buildPartial();
-    }
-
-    /**
-     * Helper for DynamicMessage.parseFrom() methods to call.  Throws
-     * {@link InvalidProtocolBufferException} instead of
-     * {@link UninitializedMessageException}.
-     */
-    private DynamicMessage buildParsed() throws InvalidProtocolBufferException {
-      if (!isInitialized()) {
-        throw newUninitializedMessageException(
-          new DynamicMessage(type, fields,
-              java.util.Arrays.copyOf(oneofCases, oneofCases.length), unknownFields))
-          .asInvalidProtocolBufferException();
-      }
-      return buildPartial();
-    }
-
-    @Override
-    public DynamicMessage buildPartial() {
-      fields.makeImmutable();
-      DynamicMessage result =
-        new DynamicMessage(type, fields,
-            java.util.Arrays.copyOf(oneofCases, oneofCases.length), unknownFields);
-      return result;
-    }
-
-    @Override
-    public Builder clone() {
-      Builder result = new Builder(type);
-      result.fields.mergeFrom(fields);
-      result.mergeUnknownFields(unknownFields);
-      System.arraycopy(oneofCases, 0, result.oneofCases, 0 , oneofCases.length);
-      return result;
-    }
-
-    @Override
-    public boolean isInitialized() {
-      return DynamicMessage.isInitialized(type, fields);
-    }
-
-    @Override
-    public Descriptor getDescriptorForType() {
-      return type;
-    }
-
-    @Override
-    public DynamicMessage getDefaultInstanceForType() {
-      return getDefaultInstance(type);
-    }
-
-    @Override
-    public Map<FieldDescriptor, Object> getAllFields() {
-      return fields.getAllFields();
-    }
-
-    @Override
-    public Builder newBuilderForField(FieldDescriptor field) {
-      verifyContainingType(field);
-
-      if (field.getJavaType() != FieldDescriptor.JavaType.MESSAGE) {
-        throw new IllegalArgumentException(
-          "newBuilderForField is only valid for fields with message type.");
-      }
-
-      return new Builder(field.getMessageType());
-    }
-
-    @Override
-    public boolean hasOneof(OneofDescriptor oneof) {
-      verifyOneofContainingType(oneof);
-      FieldDescriptor field = oneofCases[oneof.getIndex()];
-      if (field == null) {
-        return false;
-      }
-      return true;
-    }
-
-    @Override
-    public FieldDescriptor getOneofFieldDescriptor(OneofDescriptor oneof) {
-      verifyOneofContainingType(oneof);
-      return oneofCases[oneof.getIndex()];
-    }
-
-    @Override
-    public Builder clearOneof(OneofDescriptor oneof) {
-      verifyOneofContainingType(oneof);
-      FieldDescriptor field = oneofCases[oneof.getIndex()];
-      if (field != null) {
-        clearField(field);
-      }
-      return this;
-    }
-
-    @Override
-    public boolean hasField(FieldDescriptor field) {
-      verifyContainingType(field);
-      return fields.hasField(field);
-    }
-
-    @Override
-    public Object getField(FieldDescriptor field) {
-      verifyContainingType(field);
-      Object result = fields.getField(field);
-      if (result == null) {
-        if (field.isRepeated()) {
-          result = Collections.emptyList();
-        } else if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-          result = getDefaultInstance(field.getMessageType());
-        } else {
-          result = field.getDefaultValue();
-        }
-      }
-      return result;
-    }
-
-    @Override
-    public Builder setField(FieldDescriptor field, Object value) {
-      verifyContainingType(field);
-      ensureIsMutable();
-      // TODO(xiaofeng): This check should really be put in FieldSet.setField()
-      // where all other such checks are done. However, currently
-      // FieldSet.setField() permits Integer value for enum fields probably
-      // because of some internal features we support. Should figure it out
-      // and move this check to a more appropriate place.
-      if (field.getType() == FieldDescriptor.Type.ENUM) {
-        ensureEnumValueDescriptor(field, value);
-      }
-      OneofDescriptor oneofDescriptor = field.getContainingOneof();
-      if (oneofDescriptor != null) {
-        int index = oneofDescriptor.getIndex();
-        FieldDescriptor oldField = oneofCases[index];
-        if ((oldField != null) && (oldField != field)) {
-          fields.clearField(oldField);
-        }
-        oneofCases[index] = field;
-      } else if (field.getFile().getSyntax() == Descriptors.FileDescriptor.Syntax.PROTO3) {
-        if (!field.isRepeated()
-            && field.getJavaType() != FieldDescriptor.JavaType.MESSAGE
-            && value.equals(field.getDefaultValue())) {
-          // In proto3, setting a field to its default value is equivalent to clearing the field.
-          fields.clearField(field);
-          return this;
-        }
-      }
-      fields.setField(field, value);
-      return this;
-    }
-
-    @Override
-    public Builder clearField(FieldDescriptor field) {
-      verifyContainingType(field);
-      ensureIsMutable();
-      OneofDescriptor oneofDescriptor = field.getContainingOneof();
-      if (oneofDescriptor != null) {
-        int index = oneofDescriptor.getIndex();
-        if (oneofCases[index] == field) {
-          oneofCases[index] = null;
-        }
-      }
-      fields.clearField(field);
-      return this;
-    }
-
-    @Override
-    public int getRepeatedFieldCount(FieldDescriptor field) {
-      verifyContainingType(field);
-      return fields.getRepeatedFieldCount(field);
-    }
-
-    @Override
-    public Object getRepeatedField(FieldDescriptor field, int index) {
-      verifyContainingType(field);
-      return fields.getRepeatedField(field, index);
-    }
-
-    @Override
-    public Builder setRepeatedField(FieldDescriptor field, int index, Object value) {
-      verifyContainingType(field);
-      ensureIsMutable();
-      fields.setRepeatedField(field, index, value);
-      return this;
-    }
-
-    @Override
-    public Builder addRepeatedField(FieldDescriptor field, Object value) {
-      verifyContainingType(field);
-      ensureIsMutable();
-      fields.addRepeatedField(field, value);
-      return this;
-    }
-
-    @Override
-    public UnknownFieldSet getUnknownFields() {
-      return unknownFields;
-    }
-
-    @Override
-    public Builder setUnknownFields(UnknownFieldSet unknownFields) {
-      if (getDescriptorForType().getFile().getSyntax()
-          == Descriptors.FileDescriptor.Syntax.PROTO3) {
-        // Proto3 discards unknown fields.
-        return this;
-      }
-      this.unknownFields = unknownFields;
-      return this;
-    }
-
-    @Override
-    public Builder mergeUnknownFields(UnknownFieldSet unknownFields) {
-      if (getDescriptorForType().getFile().getSyntax()
-          == Descriptors.FileDescriptor.Syntax.PROTO3) {
-        // Proto3 discards unknown fields.
-        return this;
-      }
-      this.unknownFields =
-        UnknownFieldSet.newBuilder(this.unknownFields)
-                       .mergeFrom(unknownFields)
-                       .build();
-      return this;
-    }
-
-    /** Verifies that the field is a field of this message. */
-    private void verifyContainingType(FieldDescriptor field) {
-      if (field.getContainingType() != type) {
-        throw new IllegalArgumentException(
-          "FieldDescriptor does not match message type.");
-      }
-    }
-
-    /** Verifies that the oneof is an oneof of this message. */
-    private void verifyOneofContainingType(OneofDescriptor oneof) {
-      if (oneof.getContainingType() != type) {
-        throw new IllegalArgumentException(
-          "OneofDescriptor does not match message type.");
-      }
-    }
-
-    /** Verifies that the value is EnumValueDescriptor and matches Enum Type. */
-    private void ensureSingularEnumValueDescriptor(
-        FieldDescriptor field, Object value) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-      if (!(value instanceof EnumValueDescriptor)) {
-        throw new IllegalArgumentException(
-          "DynamicMessage should use EnumValueDescriptor to set Enum Value.");
-      }
-      // TODO(xiaofeng): Re-enable this check after Orgstore is fixed to not
-      // set incorrect EnumValueDescriptors.
-      // EnumDescriptor fieldType = field.getEnumType();
-      // EnumDescriptor fieldValueType = ((EnumValueDescriptor) value).getType();
-      // if (fieldType != fieldValueType) {
-      //  throw new IllegalArgumentException(String.format(
-      //      "EnumDescriptor %s of field doesn't match EnumDescriptor %s of field value",
-      //      fieldType.getFullName(), fieldValueType.getFullName()));
-      // }
-    }
-
-    /** Verifies the value for an enum field. */
-    private void ensureEnumValueDescriptor(
-        FieldDescriptor field, Object value) {
-      if (field.isRepeated()) {
-        for (Object item : (List) value) {
-          ensureSingularEnumValueDescriptor(field, item);
-        }
-      } else {
-         ensureSingularEnumValueDescriptor(field, value);
-      }
-    }
-
-    private void ensureIsMutable() {
-      if (fields.isImmutable()) {
-        fields = fields.clone();
-      }
-    }
-
-    @Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Message.Builder getFieldBuilder(FieldDescriptor field) {
-      // TODO(xiangl): need implementation for dynamic message
-      throw new UnsupportedOperationException(
-        "getFieldBuilder() called on a dynamic message type.");
-    }
-
-    @Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Message.Builder getRepeatedFieldBuilder(FieldDescriptor field,
-        int index) {
-      throw new UnsupportedOperationException(
-        "getRepeatedFieldBuilder() called on a dynamic message type.");
-    }
-  }
-}


[22/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyFieldLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyFieldLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyFieldLite.java
deleted file mode 100644
index 91b70e0..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyFieldLite.java
+++ /dev/null
@@ -1,437 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-
-/**
- * LazyFieldLite encapsulates the logic of lazily parsing message fields. It stores
- * the message in a ByteString initially and then parses it on-demand.
- *
- * LazyFieldLite is thread-compatible: concurrent reads are safe once the proto that this
- * LazyFieldLite is a part of is no longer being mutated by its Builder. However, explicit
- * synchronization is needed under read/write situations.
- *
- * When a LazyFieldLite is used in the context of a MessageLite object, its behavior is considered
- * to be immutable and none of the setter methods in its API are expected to be invoked. All of the
- * getters are expected to be thread-safe. When used in the context of a MessageLite.Builder,
- * setters can be invoked, but there is no guarantee of thread safety.
- * 
- * TODO(yatin,dweis): Consider splitting this class's functionality and put the mutable methods
- * into a separate builder class to allow us to give stronger compile-time guarantees.
- *
- * This class is internal implementation detail of the protobuf library, so you don't need to use it
- * directly.
- *
- * @author xiangl@google.com (Xiang Li)
- */
-public class LazyFieldLite {
-  private static final ExtensionRegistryLite EMPTY_REGISTRY =
-      ExtensionRegistryLite.getEmptyRegistry();
-
-  /**
-   * The value associated with the LazyFieldLite object is stored in one or more of the following
-   * three fields (delayedBytes, value, memoizedBytes). They should together be interpreted as
-   * follows.
-   * 1) delayedBytes can be non-null, while value and memoizedBytes is null. The object will be in
-   *    this state while the value for the object has not yet been parsed.
-   * 2) Both delayedBytes and value are non-null. The object transitions to this state as soon as
-   *    some caller needs to access the value (by invoking getValue()).
-   * 3) memoizedBytes is merely an optimization for calls to LazyFieldLite.toByteString() to avoid
-   *    recomputing the ByteString representation on each call. Instead, when the value is parsed
-   *    from delayedBytes, we will also assign the contents of delayedBytes to memoizedBytes (since
-   *    that is the ByteString representation of value).
-   * 4) Finally, if the LazyFieldLite was created directly with a parsed MessageLite value, then
-   *    delayedBytes will be null, and memoizedBytes will be initialized only upon the first call to
-   *    LazyFieldLite.toByteString().
-   *
-   * Given the above conditions, any caller that needs a serialized representation of this object
-   * must first check if the memoizedBytes or delayedBytes ByteString is non-null and use it
-   * directly; if both of those are null, it can look at the parsed value field. Similarly, any
-   * caller that needs a parsed value must first check if the value field is already non-null, if
-   * not it must parse the value from delayedBytes.
-   */
-
-  /**
-   * A delayed-parsed version of the contents of this field. When this field is non-null, then the
-   * "value" field is allowed to be null until the time that the value needs to be read.
-   *
-   * When delayedBytes is non-null then {@code extensionRegistry} is required to also be non-null.
-   * {@code value} and {@code memoizedBytes} will be initialized lazily.
-   */
-  private ByteString delayedBytes;
-
-  /**
-   * An {@code ExtensionRegistryLite} for parsing bytes. It is non-null on a best-effort basis. It
-   * is only guaranteed to be non-null if this message was initialized using bytes and an
-   * {@code ExtensionRegistry}. If it directly had a value set then it will be null, unless it has
-   * been merged with another {@code LazyFieldLite} that had an {@code ExtensionRegistry}.
-   */
-  private ExtensionRegistryLite extensionRegistry;
-
-  /**
-   * The parsed value. When this is null and a caller needs access to the MessageLite value, then
-   * {@code delayedBytes} will be parsed lazily at that time.
-   */
-  protected volatile MessageLite value;
-
-  /**
-   * The memoized bytes for {@code value}. This is an optimization for the toByteString() method to
-   * not have to recompute its return-value on each invocation.
-   * TODO(yatin): Figure out whether this optimization is actually necessary.
-   */
-  private volatile ByteString memoizedBytes;
-
-  /**
-   * Constructs a LazyFieldLite with bytes that will be parsed lazily.
-   */
-  public LazyFieldLite(ExtensionRegistryLite extensionRegistry, ByteString bytes) {
-    checkArguments(extensionRegistry, bytes);
-    this.extensionRegistry = extensionRegistry;
-    this.delayedBytes = bytes;
-  }
-
-  /**
-   * Constructs a LazyFieldLite with no contents, and no ability to parse extensions.
-   */
-  public LazyFieldLite() {
-  }
-
-  /**
-   * Constructs a LazyFieldLite instance with a value. The LazyFieldLite may not be able to parse
-   * the extensions in the value as it has no ExtensionRegistry.
-   */
-  public static LazyFieldLite fromValue(MessageLite value) {
-    LazyFieldLite lf = new LazyFieldLite();
-    lf.setValue(value);
-    return lf;
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (this == o) {
-      return true;
-    }
-    
-    if (!(o instanceof LazyFieldLite)) {
-      return false;
-    }
-
-    LazyFieldLite other = (LazyFieldLite) o;
-    
-    // Lazy fields do not work well with equals... If both are delayedBytes, we do not have a
-    // mechanism to deserialize them so we rely on bytes equality. Otherwise we coerce into an
-    // actual message (if necessary) and call equals on the message itself. This implies that two
-    // messages can by unequal but then be turned equal simply be invoking a getter on a lazy field.
-    MessageLite value1 = value;
-    MessageLite value2 = other.value;
-    if (value1 == null && value2 == null) {
-      return toByteString().equals(other.toByteString());
-    } else if (value1 != null && value2 != null) {
-      return value1.equals(value2);
-    } else if (value1 != null) {
-      return value1.equals(other.getValue(value1.getDefaultInstanceForType()));
-    } else {
-      return getValue(value2.getDefaultInstanceForType()).equals(value2);
-    }
-  }
-  
-  @Override
-  public int hashCode() {
-    // We can't provide a memoizable hash code for lazy fields. The byte strings may have different
-    // hash codes but evaluate to equivalent messages. And we have no facility for constructing
-    // a message here if we were not already holding a value.
-    return 1;
-  }
-  
-  /**
-   * Determines whether this LazyFieldLite instance represents the default instance of this type.
-   */
-  public boolean containsDefaultInstance() {
-    return memoizedBytes == ByteString.EMPTY
-        || value == null && (delayedBytes == null || delayedBytes == ByteString.EMPTY);
-  }
-
-  /**
-   * Clears the value state of this instance.
-   *
-   * <p>LazyField is not thread-safe for write access. Synchronizations are needed
-   * under read/write situations.
-   */
-  public void clear() {
-    // Don't clear the ExtensionRegistry. It might prove useful later on when merging in another
-    // value, but there is no guarantee that it will contain all extensions that were directly set
-    // on the values that need to be merged.
-    delayedBytes = null;
-    value = null;
-    memoizedBytes = null;
-  }
-
-  /**
-   * Overrides the contents of this LazyField.
-   *
-   * <p>LazyField is not thread-safe for write access. Synchronizations are needed
-   * under read/write situations.
-   */
-  public void set(LazyFieldLite other) {
-    this.delayedBytes = other.delayedBytes;
-    this.value = other.value;
-    this.memoizedBytes = other.memoizedBytes;
-    // If the other LazyFieldLite was created by directly setting the value rather than first by
-    // parsing, then it will not have an extensionRegistry. In this case we hold on to the existing
-    // extensionRegistry, which has no guarantees that it has all the extensions that will be
-    // directly set on the value.
-    if (other.extensionRegistry != null) {
-      this.extensionRegistry = other.extensionRegistry;
-    }
-  }
-
-  /**
-   * Returns message instance. It may do some thread-safe delayed parsing of bytes.
-   *
-   * @param defaultInstance its message's default instance. It's also used to get parser for the
-   * message type.
-   */
-  public MessageLite getValue(MessageLite defaultInstance) {
-    ensureInitialized(defaultInstance);
-    return value;
-  }
-
-  /**
-   * Sets the value of the instance and returns the old value without delay parsing anything.
-   *
-   * <p>LazyField is not thread-safe for write access. Synchronizations are needed
-   * under read/write situations.
-   */
-  public MessageLite setValue(MessageLite value) {
-    MessageLite originalValue = this.value;
-    this.delayedBytes = null;
-    this.memoizedBytes = null;
-    this.value = value;
-    return originalValue;
-  }
-
-  /**
-   * Merges another instance's contents. In some cases may drop some extensions if both fields
-   * contain data. If the other field has an {@code ExtensionRegistry} but this does not, then this
-   * field will copy over that {@code ExtensionRegistry}.
-   *
-   * <p>LazyField is not thread-safe for write access. Synchronizations are needed
-   * under read/write situations.
-   */
-  public void merge(LazyFieldLite other) {
-    if (other.containsDefaultInstance()) {
-      return;
-    }
-
-    if (this.containsDefaultInstance()) {
-      set(other);
-      return;
-    }
-
-    // If the other field has an extension registry but this does not, copy over the other extension
-    // registry.
-    if (this.extensionRegistry == null) {
-      this.extensionRegistry = other.extensionRegistry;
-    }
-
-    // In the case that both of them are not parsed we simply concatenate the bytes to save time. In
-    // the (probably rare) case that they have different extension registries there is a chance that
-    // some of the extensions may be dropped, but the tradeoff of making this operation fast seems
-    // to outway the benefits of combining the extension registries, which is not normally done for
-    // lite protos anyways.
-    if (this.delayedBytes != null && other.delayedBytes != null) {
-      this.delayedBytes = this.delayedBytes.concat(other.delayedBytes);
-      return;
-    }
-
-    // At least one is parsed and both contain data. We won't drop any extensions here directly, but
-    // in the case that the extension registries are not the same then we might in the future if we
-    // need to serialze and parse a message again.
-    if (this.value == null && other.value != null) {
-      setValue(mergeValueAndBytes(other.value, this.delayedBytes, this.extensionRegistry));
-      return;
-    } else if (this.value != null && other.value == null) {
-      setValue(mergeValueAndBytes(this.value, other.delayedBytes, other.extensionRegistry));
-      return;
-    }
-
-    // At this point we have two fully parsed messages.
-    setValue(this.value.toBuilder().mergeFrom(other.value).build());
-  }
-  
-  /**
-   * Merges another instance's contents from a stream.
-   *
-   * <p>LazyField is not thread-safe for write access. Synchronizations are needed
-   * under read/write situations.
-   */
-  public void mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-      throws IOException {
-    if (this.containsDefaultInstance()) {
-      setByteString(input.readBytes(), extensionRegistry);
-      return;
-    }
-
-    // If the other field has an extension registry but this does not, copy over the other extension
-    // registry.
-    if (this.extensionRegistry == null) {
-      this.extensionRegistry = extensionRegistry;
-    }
-
-    // In the case that both of them are not parsed we simply concatenate the bytes to save time. In
-    // the (probably rare) case that they have different extension registries there is a chance that
-    // some of the extensions may be dropped, but the tradeoff of making this operation fast seems
-    // to outway the benefits of combining the extension registries, which is not normally done for
-    // lite protos anyways.
-    if (this.delayedBytes != null) {
-      setByteString(this.delayedBytes.concat(input.readBytes()), this.extensionRegistry);
-      return;
-    }
-
-    // We are parsed and both contain data. We won't drop any extensions here directly, but in the
-    // case that the extension registries are not the same then we might in the future if we
-    // need to serialize and parse a message again.
-    try {
-      setValue(value.toBuilder().mergeFrom(input, extensionRegistry).build());
-    } catch (InvalidProtocolBufferException e) {
-      // Nothing is logged and no exceptions are thrown. Clients will be unaware that a proto
-      // was invalid.
-    }
-  }
-
-  private static MessageLite mergeValueAndBytes(
-      MessageLite value, ByteString otherBytes, ExtensionRegistryLite extensionRegistry) {
-    try {
-      return value.toBuilder().mergeFrom(otherBytes, extensionRegistry).build();
-    } catch (InvalidProtocolBufferException e) {
-      // Nothing is logged and no exceptions are thrown. Clients will be unaware that a proto
-      // was invalid.
-      return value;
-    }
-  }
-
-  /**
-   * Sets this field with bytes to delay-parse.
-   */
-  public void setByteString(ByteString bytes, ExtensionRegistryLite extensionRegistry) {
-    checkArguments(extensionRegistry, bytes);
-    this.delayedBytes = bytes;
-    this.extensionRegistry = extensionRegistry;
-    this.value = null;
-    this.memoizedBytes = null;
-  }
-
-  /**
-   * Due to the optional field can be duplicated at the end of serialized
-   * bytes, which will make the serialized size changed after LazyField
-   * parsed. Be careful when using this method.
-   */
-  public int getSerializedSize() {
-    // We *must* return delayed bytes size if it was ever set because the dependent messages may
-    // have memoized serialized size based off of it.
-    if (memoizedBytes != null) {
-      return memoizedBytes.size();
-    } else if (delayedBytes != null) {
-      return delayedBytes.size();
-    } else if (value != null) {
-      return value.getSerializedSize();
-    } else {
-      return 0;
-    }
-  }
-
-  /**
-   * Returns a BytesString for this field in a thread-safe way.
-   */
-  public ByteString toByteString() {
-    if (memoizedBytes != null) {
-      return memoizedBytes;
-    }
-    // We *must* return delayed bytes if it was set because the dependent messages may have
-    // memoized serialized size based off of it.
-    if (delayedBytes != null) {
-      return delayedBytes;
-    }
-    synchronized (this) {
-      if (memoizedBytes != null) {
-        return memoizedBytes;
-      }
-      if (value == null) {
-        memoizedBytes = ByteString.EMPTY;
-      } else {
-        memoizedBytes = value.toByteString();
-      }
-      return memoizedBytes;
-    }
-  }
-
-  /**
-   * Might lazily parse the bytes that were previously passed in. Is thread-safe.
-   */
-  protected void ensureInitialized(MessageLite defaultInstance) {
-    if (value != null) {
-      return;
-    }
-    synchronized (this) {
-      if (value != null) {
-        return;
-      }
-      try {
-        if (delayedBytes != null) {
-          // The extensionRegistry shouldn't be null here since we have delayedBytes.
-          MessageLite parsedValue = defaultInstance.getParserForType()
-              .parseFrom(delayedBytes, extensionRegistry);
-          this.value = parsedValue;
-          this.memoizedBytes = delayedBytes;
-        } else {
-          this.value = defaultInstance;
-          this.memoizedBytes = ByteString.EMPTY;
-        }
-      } catch (InvalidProtocolBufferException e) {
-        // Nothing is logged and no exceptions are thrown. Clients will be unaware that this proto
-        // was invalid.
-        this.value = defaultInstance;
-        this.memoizedBytes = ByteString.EMPTY;
-      }
-    }
-  }
-
-
-  private static void checkArguments(ExtensionRegistryLite extensionRegistry, ByteString bytes) {
-    if (extensionRegistry == null) {
-      throw new NullPointerException("found null ExtensionRegistry");
-    }
-    if (bytes == null) {
-      throw new NullPointerException("found null ByteString");
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyStringArrayList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyStringArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyStringArrayList.java
deleted file mode 100644
index 1f99489..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyStringArrayList.java
+++ /dev/null
@@ -1,423 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.AbstractList;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.List;
-import java.util.RandomAccess;
-
-/**
- * An implementation of {@link LazyStringList} that wraps an ArrayList. Each
- * element is one of String, ByteString, or byte[]. It caches the last one
- * requested which is most likely the one needed next. This minimizes memory
- * usage while satisfying the most common use cases.
- * <p>
- * <strong>Note that this implementation is not synchronized.</strong>
- * If multiple threads access an <tt>ArrayList</tt> instance concurrently,
- * and at least one of the threads modifies the list structurally, it
- * <i>must</i> be synchronized externally.  (A structural modification is
- * any operation that adds or deletes one or more elements, or explicitly
- * resizes the backing array; merely setting the value of an element is not
- * a structural modification.)  This is typically accomplished by
- * synchronizing on some object that naturally encapsulates the list.
- * <p>
- * If the implementation is accessed via concurrent reads, this is thread safe.
- * Conversions are done in a thread safe manner. It's possible that the
- * conversion may happen more than once if two threads attempt to access the
- * same element and the modifications were not visible to each other, but this
- * will not result in any corruption of the list or change in behavior other
- * than performance.
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public class LazyStringArrayList extends AbstractProtobufList<String>
-    implements LazyStringList, RandomAccess {
-
-  private static final LazyStringArrayList EMPTY_LIST = new LazyStringArrayList();
-  static {
-    EMPTY_LIST.makeImmutable();
-  }
-  
-  static LazyStringArrayList emptyList() {
-    return EMPTY_LIST;
-  }
-
-  // For compatibility with older runtimes.
-  public static final LazyStringList EMPTY = EMPTY_LIST;
-
-  private final List<Object> list;
-
-  public LazyStringArrayList() {
-    this(DEFAULT_CAPACITY);
-  }
-
-  public LazyStringArrayList(int intialCapacity) {
-    this(new ArrayList<Object>(intialCapacity));
-  }
-
-  public LazyStringArrayList(LazyStringList from) {
-    list = new ArrayList<Object>(from.size());
-    addAll(from);
-  }
-
-  public LazyStringArrayList(List<String> from) {
-    this(new ArrayList<Object>(from));
-  }
-  
-  private LazyStringArrayList(ArrayList<Object> list) {
-    this.list = list;
-  }
-
-  @Override
-  public LazyStringArrayList mutableCopyWithCapacity(int capacity) {
-    if (capacity < size()) {
-      throw new IllegalArgumentException();
-    }
-    ArrayList<Object> newList = new ArrayList<Object>(capacity);
-    newList.addAll(list);
-    return new LazyStringArrayList(newList);
-  }
-
-  @Override
-  public String get(int index) {
-    Object o = list.get(index);
-    if (o instanceof String) {
-      return (String) o;
-    } else if (o instanceof ByteString) {
-      ByteString bs = (ByteString) o;
-      String s = bs.toStringUtf8();
-      if (bs.isValidUtf8()) {
-        list.set(index, s);
-      }
-      return s;
-    } else {
-      byte[] ba = (byte[]) o;
-      String s = Internal.toStringUtf8(ba);
-      if (Internal.isValidUtf8(ba)) {
-        list.set(index, s);
-      }
-      return s;
-    }
-  }
-
-  @Override
-  public int size() {
-    return list.size();
-  }
-
-  @Override
-  public String set(int index, String s) {
-    ensureIsMutable();
-    Object o = list.set(index, s);
-    return asString(o);
-  }
-
-  @Override
-  public void add(int index, String element) {
-    ensureIsMutable();
-    list.add(index, element);
-    modCount++;
-  }
-  
-  private void add(int index, ByteString element) {
-    ensureIsMutable();
-    list.add(index, element);
-    modCount++;
-  }
-  
-  private void add(int index, byte[] element) {
-    ensureIsMutable();
-    list.add(index, element);
-    modCount++;
-  }
-
-  @Override
-  public boolean addAll(Collection<? extends String> c) {
-    // The default implementation of AbstractCollection.addAll(Collection)
-    // delegates to add(Object). This implementation instead delegates to
-    // addAll(int, Collection), which makes a special case for Collections
-    // which are instances of LazyStringList.
-    return addAll(size(), c);
-  }
-
-  @Override
-  public boolean addAll(int index, Collection<? extends String> c) {
-    ensureIsMutable();
-    // When copying from another LazyStringList, directly copy the underlying
-    // elements rather than forcing each element to be decoded to a String.
-    Collection<?> collection = c instanceof LazyStringList
-        ? ((LazyStringList) c).getUnderlyingElements() : c;
-    boolean ret = list.addAll(index, collection);
-    modCount++;
-    return ret;
-  }
-
-  @Override
-  public boolean addAllByteString(Collection<? extends ByteString> values) {
-    ensureIsMutable();
-    boolean ret = list.addAll(values);
-    modCount++;
-    return ret;
-  }
-
-  @Override
-  public boolean addAllByteArray(Collection<byte[]> c) {
-    ensureIsMutable();
-    boolean ret = list.addAll(c);
-    modCount++;
-    return ret;
-  }
-
-  @Override
-  public String remove(int index) {
-    ensureIsMutable();
-    Object o = list.remove(index);
-    modCount++;
-    return asString(o);
-  }
-
-  @Override
-  public void clear() {
-    ensureIsMutable();
-    list.clear();
-    modCount++;
-  }
-
-  @Override
-  public void add(ByteString element) {
-    ensureIsMutable();
-    list.add(element);
-    modCount++;
-  }
-  
-  @Override
-  public void add(byte[] element) {
-    ensureIsMutable();
-    list.add(element);
-    modCount++;
-  }
-
-  @Override
-  public Object getRaw(int index) {
-    return list.get(index);
-  }
-  
-  @Override
-  public ByteString getByteString(int index) {
-    Object o = list.get(index);
-    ByteString b = asByteString(o);
-    if (b != o) {
-      list.set(index, b);
-    }
-    return b;
-  }
-  
-  @Override
-  public byte[] getByteArray(int index) {
-    Object o = list.get(index);
-    byte[] b = asByteArray(o);
-    if (b != o) {
-      list.set(index, b);
-    }
-    return b;
-  }
-
-  @Override
-  public void set(int index, ByteString s) {
-    setAndReturn(index, s);
-  }
-  
-  private Object setAndReturn(int index, ByteString s) {
-    ensureIsMutable();
-    return list.set(index, s);
-  }
-
-  @Override
-  public void set(int index, byte[] s) {
-    setAndReturn(index, s);
-  }
-  
-  private Object setAndReturn(int index, byte[] s) {
-    ensureIsMutable();
-    return list.set(index, s);
-  }
-
-  private static String asString(Object o) {
-    if (o instanceof String) {
-      return (String) o;
-    } else if (o instanceof ByteString) {
-      return ((ByteString) o).toStringUtf8();
-    } else {
-      return Internal.toStringUtf8((byte[]) o);
-    }
-  }
-  
-  private static ByteString asByteString(Object o) {
-    if (o instanceof ByteString) {
-      return (ByteString) o;
-    } else if (o instanceof String) {
-      return ByteString.copyFromUtf8((String) o);
-    } else {
-      return ByteString.copyFrom((byte[]) o);
-    }
-  }
-  
-  private static byte[] asByteArray(Object o) {
-    if (o instanceof byte[]) {
-      return (byte[]) o;
-    } else if (o instanceof String) {
-      return Internal.toByteArray((String) o);
-    } else {
-      return ((ByteString) o).toByteArray();
-    }
-  }
-
-  @Override
-  public List<?> getUnderlyingElements() {
-    return Collections.unmodifiableList(list);
-  }
-
-  @Override
-  public void mergeFrom(LazyStringList other) {
-    ensureIsMutable();
-    for (Object o : other.getUnderlyingElements()) {
-      if (o instanceof byte[]) {
-        byte[] b = (byte[]) o;
-        // Byte array's content is mutable so they should be copied rather than
-        // shared when merging from one message to another.
-        list.add(Arrays.copyOf(b, b.length));
-      } else {
-        list.add(o);
-      }
-    }
-  }
-
-  private static class ByteArrayListView extends AbstractList<byte[]>
-      implements RandomAccess {
-    private final LazyStringArrayList list;
-    
-    ByteArrayListView(LazyStringArrayList list) {
-      this.list = list;
-    }
-    
-    @Override
-    public byte[] get(int index) {
-      return list.getByteArray(index);
-    }
-
-    @Override
-    public int size() {
-      return list.size();
-    }
-
-    @Override
-    public byte[] set(int index, byte[] s) {
-      Object o = list.setAndReturn(index, s);
-      modCount++;
-      return asByteArray(o);
-    }
-
-    @Override
-    public void add(int index, byte[] s) {
-      list.add(index, s);
-      modCount++;
-    }
-
-    @Override
-    public byte[] remove(int index) {
-      Object o = list.remove(index);
-      modCount++;
-      return asByteArray(o);
-    }
-  }
-  
-  @Override
-  public List<byte[]> asByteArrayList() {
-    return new ByteArrayListView(this);
-  }
-
-  private static class ByteStringListView extends AbstractList<ByteString>
-      implements RandomAccess {
-    private final LazyStringArrayList list;
-
-    ByteStringListView(LazyStringArrayList list) {
-      this.list = list;
-    }
-
-    @Override
-    public ByteString get(int index) {
-      return list.getByteString(index);
-    }
-
-    @Override
-    public int size() {
-      return list.size();
-    }
-
-    @Override
-    public ByteString set(int index, ByteString s) {
-      Object o = list.setAndReturn(index, s);
-      modCount++;
-      return asByteString(o);
-    }
-
-    @Override
-    public void add(int index, ByteString s) {
-      list.add(index, s);
-      modCount++;
-    }
-
-    @Override
-    public ByteString remove(int index) {
-      Object o = list.remove(index);
-      modCount++;
-      return asByteString(o);
-    }
-  }
-
-  @Override
-  public List<ByteString> asByteStringList() {
-    return new ByteStringListView(this);
-  }
-
-  @Override
-  public LazyStringList getUnmodifiableView() {
-    if (isModifiable()) {
-      return new UnmodifiableLazyStringList(this);
-    }
-    return this;
-  }
-
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyStringList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyStringList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyStringList.java
deleted file mode 100644
index 5412763..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyStringList.java
+++ /dev/null
@@ -1,174 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.Collection;
-import java.util.List;
-
-/**
- * An interface extending {@code List<String>} that also provides access to the
- * items of the list as UTF8-encoded ByteString or byte[] objects. This is
- * used by the protocol buffer implementation to support lazily converting bytes
- * parsed over the wire to String objects until needed and also increases the
- * efficiency of serialization if the String was never requested as the
- * ByteString or byte[] is already cached. The ByteString methods are used in
- * immutable API only and byte[] methods used in mutable API only for they use
- * different representations for string/bytes fields.
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public interface LazyStringList extends ProtocolStringList {
-
-  /**
-   * Returns the element at the specified position in this list as a ByteString.
-   *
-   * @param index index of the element to return
-   * @return the element at the specified position in this list
-   * @throws IndexOutOfBoundsException if the index is out of range
-   *         ({@code index < 0 || index >= size()})
-   */
-  ByteString getByteString(int index);
-
-  /**
-   * Returns the element at the specified position in this list as an Object
-   * that will either be a String or a ByteString.
-   *
-   * @param index index of the element to return
-   * @return the element at the specified position in this list
-   * @throws IndexOutOfBoundsException if the index is out of range
-   *         ({@code index < 0 || index >= size()})
-   */
-  Object getRaw(int index);
-
-  /**
-   * Returns the element at the specified position in this list as byte[].
-   *
-   * @param index index of the element to return
-   * @return the element at the specified position in this list
-   * @throws IndexOutOfBoundsException if the index is out of range
-   *         ({@code index < 0 || index >= size()})
-   */
-  byte[] getByteArray(int index);
-
-  /**
-   * Appends the specified element to the end of this list (optional
-   * operation).
-   *
-   * @param element element to be appended to this list
-   * @throws UnsupportedOperationException if the <tt>add</tt> operation
-   *         is not supported by this list
-   */
-  void add(ByteString element);
-
-  /**
-   * Appends the specified element to the end of this list (optional
-   * operation).
-   *
-   * @param element element to be appended to this list
-   * @throws UnsupportedOperationException if the <tt>add</tt> operation
-   *         is not supported by this list
-   */
-  void add(byte[] element);
-
-  /**
-   * Replaces the element at the specified position in this list with the
-   * specified element (optional operation).
-   *
-   * @param index index of the element to replace
-   * @param element the element to be stored at the specified position
-   * @throws UnsupportedOperationException if the <tt>set</tt> operation
-   *         is not supported by this list
-   *         IndexOutOfBoundsException if the index is out of range
-   *         ({@code index < 0 || index >= size()})
-   */
-  void set(int index, ByteString element);
-  
-  /**
-   * Replaces the element at the specified position in this list with the
-   * specified element (optional operation).
-   *
-   * @param index index of the element to replace
-   * @param element the element to be stored at the specified position
-   * @throws UnsupportedOperationException if the <tt>set</tt> operation
-   *         is not supported by this list
-   *         IndexOutOfBoundsException if the index is out of range
-   *         ({@code index < 0 || index >= size()})
-   */
-  void set(int index, byte[] element);
-
-  /**
-   * Appends all elements in the specified ByteString collection to the end of
-   * this list.
-   *
-   * @param c collection whose elements are to be added to this list
-   * @return true if this list changed as a result of the call
-   * @throws UnsupportedOperationException if the <tt>addAllByteString</tt>
-   *         operation is not supported by this list
-   */
-  boolean addAllByteString(Collection<? extends ByteString> c);
-
-  /**
-   * Appends all elements in the specified byte[] collection to the end of
-   * this list.
-   *
-   * @param c collection whose elements are to be added to this list
-   * @return true if this list changed as a result of the call
-   * @throws UnsupportedOperationException if the <tt>addAllByteArray</tt>
-   *         operation is not supported by this list
-   */
-  boolean addAllByteArray(Collection<byte[]> c);
-
-  /**
-   * Returns an unmodifiable List of the underlying elements, each of which is
-   * either a {@code String} or its equivalent UTF-8 encoded {@code ByteString}
-   * or byte[]. It is an error for the caller to modify the returned
-   * List, and attempting to do so will result in an
-   * {@link UnsupportedOperationException}.
-   */
-  List<?> getUnderlyingElements();
-
-  /**
-   * Merges all elements from another LazyStringList into this one. This method
-   * differs from {@link #addAll(Collection)} on that underlying byte arrays are
-   * copied instead of reference shared. Immutable API doesn't need to use this
-   * method as byte[] is not used there at all.
-   */
-  void mergeFrom(LazyStringList other);
-
-  /**
-   * Returns a mutable view of this list. Changes to the view will be made into
-   * the original list. This method is used in mutable API only.
-   */
-  List<byte[]> asByteArrayList();
-
-  /** Returns an unmodifiable view of the list. */
-  LazyStringList getUnmodifiableView();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ListValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ListValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ListValue.java
deleted file mode 100644
index 4ddc71b..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ListValue.java
+++ /dev/null
@@ -1,814 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/struct.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * `ListValue` is a wrapper around a repeated field of values.
- * The JSON representation for `ListValue` is JSON array.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.ListValue}
- */
-public  final class ListValue extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.ListValue)
-    ListValueOrBuilder {
-  // Use ListValue.newBuilder() to construct.
-  private ListValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private ListValue() {
-    values_ = java.util.Collections.emptyList();
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private ListValue(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-              values_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>();
-              mutable_bitField0_ |= 0x00000001;
-            }
-            values_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.parser(), extensionRegistry));
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-        values_ = java.util.Collections.unmodifiableList(values_);
-      }
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_ListValue_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_ListValue_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.Builder.class);
-  }
-
-  public static final int VALUES_FIELD_NUMBER = 1;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> values_;
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> getValuesList() {
-    return values_;
-  }
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder> 
-      getValuesOrBuilderList() {
-    return values_;
-  }
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  public int getValuesCount() {
-    return values_.size();
-  }
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getValues(int index) {
-    return values_.get(index);
-  }
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder getValuesOrBuilder(
-      int index) {
-    return values_.get(index);
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    for (int i = 0; i < values_.size(); i++) {
-      output.writeMessage(1, values_.get(i));
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    for (int i = 0; i < values_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(1, values_.get(i));
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) obj;
-
-    boolean result = true;
-    result = result && getValuesList()
-        .equals(other.getValuesList());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    if (getValuesCount() > 0) {
-      hash = (37 * hash) + VALUES_FIELD_NUMBER;
-      hash = (53 * hash) + getValuesList().hashCode();
-    }
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * `ListValue` is a wrapper around a repeated field of values.
-   * The JSON representation for `ListValue` is JSON array.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.ListValue}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.ListValue)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_ListValue_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_ListValue_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-        getValuesFieldBuilder();
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      if (valuesBuilder_ == null) {
-        values_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000001);
-      } else {
-        valuesBuilder_.clear();
-      }
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_ListValue_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue(this);
-      int from_bitField0_ = bitField0_;
-      if (valuesBuilder_ == null) {
-        if (((bitField0_ & 0x00000001) == 0x00000001)) {
-          values_ = java.util.Collections.unmodifiableList(values_);
-          bitField0_ = (bitField0_ & ~0x00000001);
-        }
-        result.values_ = values_;
-      } else {
-        result.values_ = valuesBuilder_.build();
-      }
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance()) return this;
-      if (valuesBuilder_ == null) {
-        if (!other.values_.isEmpty()) {
-          if (values_.isEmpty()) {
-            values_ = other.values_;
-            bitField0_ = (bitField0_ & ~0x00000001);
-          } else {
-            ensureValuesIsMutable();
-            values_.addAll(other.values_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.values_.isEmpty()) {
-          if (valuesBuilder_.isEmpty()) {
-            valuesBuilder_.dispose();
-            valuesBuilder_ = null;
-            values_ = other.values_;
-            bitField0_ = (bitField0_ & ~0x00000001);
-            valuesBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getValuesFieldBuilder() : null;
-          } else {
-            valuesBuilder_.addAllMessages(other.values_);
-          }
-        }
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> values_ =
-      java.util.Collections.emptyList();
-    private void ensureValuesIsMutable() {
-      if (!((bitField0_ & 0x00000001) == 0x00000001)) {
-        values_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>(values_);
-        bitField0_ |= 0x00000001;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Value, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder> valuesBuilder_;
-
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> getValuesList() {
-      if (valuesBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(values_);
-      } else {
-        return valuesBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public int getValuesCount() {
-      if (valuesBuilder_ == null) {
-        return values_.size();
-      } else {
-        return valuesBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getValues(int index) {
-      if (valuesBuilder_ == null) {
-        return values_.get(index);
-      } else {
-        return valuesBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder setValues(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value value) {
-      if (valuesBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureValuesIsMutable();
-        values_.set(index, value);
-        onChanged();
-      } else {
-        valuesBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder setValues(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder builderForValue) {
-      if (valuesBuilder_ == null) {
-        ensureValuesIsMutable();
-        values_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        valuesBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder addValues(org.apache.hadoop.hbase.shaded.com.google.protobuf.Value value) {
-      if (valuesBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureValuesIsMutable();
-        values_.add(value);
-        onChanged();
-      } else {
-        valuesBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder addValues(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value value) {
-      if (valuesBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureValuesIsMutable();
-        values_.add(index, value);
-        onChanged();
-      } else {
-        valuesBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder addValues(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder builderForValue) {
-      if (valuesBuilder_ == null) {
-        ensureValuesIsMutable();
-        values_.add(builderForValue.build());
-        onChanged();
-      } else {
-        valuesBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder addValues(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder builderForValue) {
-      if (valuesBuilder_ == null) {
-        ensureValuesIsMutable();
-        values_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        valuesBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder addAllValues(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> values) {
-      if (valuesBuilder_ == null) {
-        ensureValuesIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, values_);
-        onChanged();
-      } else {
-        valuesBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder clearValues() {
-      if (valuesBuilder_ == null) {
-        values_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000001);
-        onChanged();
-      } else {
-        valuesBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public Builder removeValues(int index) {
-      if (valuesBuilder_ == null) {
-        ensureValuesIsMutable();
-        values_.remove(index);
-        onChanged();
-      } else {
-        valuesBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder getValuesBuilder(
-        int index) {
-      return getValuesFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder getValuesOrBuilder(
-        int index) {
-      if (valuesBuilder_ == null) {
-        return values_.get(index);  } else {
-        return valuesBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder> 
-         getValuesOrBuilderList() {
-      if (valuesBuilder_ != null) {
-        return valuesBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(values_);
-      }
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder addValuesBuilder() {
-      return getValuesFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder addValuesBuilder(
-        int index) {
-      return getValuesFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Repeated field of dynamically typed values.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Value values = 1;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder> 
-         getValuesBuilderList() {
-      return getValuesFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Value, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder> 
-        getValuesFieldBuilder() {
-      if (valuesBuilder_ == null) {
-        valuesBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Value, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder>(
-                values_,
-                ((bitField0_ & 0x00000001) == 0x00000001),
-                getParentForChildren(),
-                isClean());
-        values_ = null;
-      }
-      return valuesBuilder_;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.ListValue)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.ListValue)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ListValue>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<ListValue>() {
-    public ListValue parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new ListValue(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ListValue> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<ListValue> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ListValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ListValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ListValueOrBuilder.java
deleted file mode 100644
index 77b3c19..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ListValueOrBuilder.java
+++ /dev/null
@@ -1,53 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/struct.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface ListValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.ListValue)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> 
-      getValuesList();
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getValues(int index);
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  int getValuesCount();
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder> 
-      getValuesOrBuilderList();
-  /**
-   * <pre>
-   * Repeated field of dynamically typed values.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Value values = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder getValuesOrBuilder(
-      int index);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LongArrayList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LongArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LongArrayList.java
deleted file mode 100644
index 39d14b2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LongArrayList.java
+++ /dev/null
@@ -1,272 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.LongList;
-
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.RandomAccess;
-
-/**
- * An implementation of {@link LongList} on top of a primitive array.
- *
- * @author dweis@google.com (Daniel Weis)
- */
-final class LongArrayList
-    extends AbstractProtobufList<Long>
-    implements LongList, RandomAccess {
-
-  private static final LongArrayList EMPTY_LIST = new LongArrayList();
-  static {
-    EMPTY_LIST.makeImmutable();
-  }
-
-  public static LongArrayList emptyList() {
-    return EMPTY_LIST;
-  }
-
-  /**
-   * The backing store for the list.
-   */
-  private long[] array;
-
-  /**
-   * The size of the list distinct from the length of the array. That is, it is the number of
-   * elements set in the list.
-   */
-  private int size;
-
-  /**
-   * Constructs a new mutable {@code LongArrayList} with default capacity.
-   */
-  LongArrayList() {
-    this(new long[DEFAULT_CAPACITY], 0);
-  }
-
-  /**
-   * Constructs a new mutable {@code LongArrayList}
-   * containing the same elements as {@code other}.
-   */
-  private LongArrayList(long[] other, int size) {
-    array = other;
-    this.size = size;
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (this == o) {
-      return true;
-    }
-    if (!(o instanceof LongArrayList)) {
-      return super.equals(o);
-    }
-    LongArrayList other = (LongArrayList) o;
-    if (size != other.size) {
-      return false;
-    }
-
-    final long[] arr = other.array;
-    for (int i = 0; i < size; i++) {
-      if (array[i] != arr[i]) {
-        return false;
-      }
-    }
-
-    return true;
-  }
-
-  @Override
-  public int hashCode() {
-    int result = 1;
-    for (int i = 0; i < size; i++) {
-      result = (31 * result) + Internal.hashLong(array[i]);
-    }
-    return result;
-  }
-
-  @Override
-  public LongList mutableCopyWithCapacity(int capacity) {
-    if (capacity < size) {
-      throw new IllegalArgumentException();
-    }
-    return new LongArrayList(Arrays.copyOf(array, capacity), size);
-  }
-
-  @Override
-  public Long get(int index) {
-    return getLong(index);
-  }
-
-  @Override
-  public long getLong(int index) {
-    ensureIndexInRange(index);
-    return array[index];
-  }
-
-  @Override
-  public int size() {
-    return size;
-  }
-
-  @Override
-  public Long set(int index, Long element) {
-    return setLong(index, element);
-  }
-
-  @Override
-  public long setLong(int index, long element) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    long previousValue = array[index];
-    array[index] = element;
-    return previousValue;
-  }
-
-  @Override
-  public void add(int index, Long element) {
-    addLong(index, element);
-  }
-
-  /**
-   * Like {@link #add(Long)} but more efficient in that it doesn't box the element.
-   */
-  @Override
-  public void addLong(long element) {
-    addLong(size, element);
-  }
-
-  /**
-   * Like {@link #add(int, Long)} but more efficient in that it doesn't box the element.
-   */
-  private void addLong(int index, long element) {
-    ensureIsMutable();
-    if (index < 0 || index > size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-
-    if (size < array.length) {
-      // Shift everything over to make room
-      System.arraycopy(array, index, array, index + 1, size - index);
-    } else {
-      // Resize to 1.5x the size
-      int length = ((size * 3) / 2) + 1;
-      long[] newArray = new long[length];
-
-      // Copy the first part directly
-      System.arraycopy(array, 0, newArray, 0, index);
-
-      // Copy the rest shifted over by one to make room
-      System.arraycopy(array, index, newArray, index + 1, size - index);
-      array = newArray;
-    }
-
-    array[index] = element;
-    size++;
-    modCount++;
-  }
-
-  @Override
-  public boolean addAll(Collection<? extends Long> collection) {
-    ensureIsMutable();
-
-    if (collection == null) {
-      throw new NullPointerException();
-    }
-
-    // We specialize when adding another LongArrayList to avoid boxing elements.
-    if (!(collection instanceof LongArrayList)) {
-      return super.addAll(collection);
-    }
-
-    LongArrayList list = (LongArrayList) collection;
-    if (list.size == 0) {
-      return false;
-    }
-
-    int overflow = Integer.MAX_VALUE - size;
-    if (overflow < list.size) {
-      // We can't actually represent a list this large.
-      throw new OutOfMemoryError();
-    }
-
-    int newSize = size + list.size;
-    if (newSize > array.length) {
-      array = Arrays.copyOf(array, newSize);
-    }
-
-    System.arraycopy(list.array, 0, array, size, list.size);
-    size = newSize;
-    modCount++;
-    return true;
-  }
-
-  @Override
-  public boolean remove(Object o) {
-    ensureIsMutable();
-    for (int i = 0; i < size; i++) {
-      if (o.equals(array[i])) {
-        System.arraycopy(array, i + 1, array, i, size - i);
-        size--;
-        modCount++;
-        return true;
-      }
-    }
-    return false;
-  }
-
-  @Override
-  public Long remove(int index) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    long value = array[index];
-    System.arraycopy(array, index + 1, array, index, size - index);
-    size--;
-    modCount++;
-    return value;
-  }
-
-  /**
-   * Ensures that the provided {@code index} is within the range of {@code [0, size]}. Throws an
-   * {@link IndexOutOfBoundsException} if it is not.
-   *
-   * @param index the index to verify is in range
-   */
-  private void ensureIndexInRange(int index) {
-    if (index < 0 || index >= size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-  }
-
-  private String makeOutOfBoundsExceptionMessage(int index) {
-    return "Index:" + index + ", Size:" + size;
-  }
-}


[49/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationProtos.java
----------------------------------------------------------------------
diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationProtos.java
deleted file mode 100644
index 61b47ff..0000000
--- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationProtos.java
+++ /dev/null
@@ -1,1277 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: ColumnAggregationProtocol.proto
-
-package org.apache.hadoop.hbase.coprocessor.protobuf.generated;
-
-public final class ColumnAggregationProtos {
-  private ColumnAggregationProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface SumRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required bytes family = 1;
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    boolean hasFamily();
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    com.google.protobuf.ByteString getFamily();
-
-    // optional bytes qualifier = 2;
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    boolean hasQualifier();
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    com.google.protobuf.ByteString getQualifier();
-  }
-  /**
-   * Protobuf type {@code SumRequest}
-   */
-  public static final class SumRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements SumRequestOrBuilder {
-    // Use SumRequest.newBuilder() to construct.
-    private SumRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private SumRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final SumRequest defaultInstance;
-    public static SumRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public SumRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private SumRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              family_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              bitField0_ |= 0x00000002;
-              qualifier_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<SumRequest> PARSER =
-        new com.google.protobuf.AbstractParser<SumRequest>() {
-      public SumRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new SumRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<SumRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required bytes family = 1;
-    public static final int FAMILY_FIELD_NUMBER = 1;
-    private com.google.protobuf.ByteString family_;
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    public boolean hasFamily() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    public com.google.protobuf.ByteString getFamily() {
-      return family_;
-    }
-
-    // optional bytes qualifier = 2;
-    public static final int QUALIFIER_FIELD_NUMBER = 2;
-    private com.google.protobuf.ByteString qualifier_;
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    public boolean hasQualifier() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    public com.google.protobuf.ByteString getQualifier() {
-      return qualifier_;
-    }
-
-    private void initFields() {
-      family_ = com.google.protobuf.ByteString.EMPTY;
-      qualifier_ = com.google.protobuf.ByteString.EMPTY;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasFamily()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, family_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeBytes(2, qualifier_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, family_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(2, qualifier_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest) obj;
-
-      boolean result = true;
-      result = result && (hasFamily() == other.hasFamily());
-      if (hasFamily()) {
-        result = result && getFamily()
-            .equals(other.getFamily());
-      }
-      result = result && (hasQualifier() == other.hasQualifier());
-      if (hasQualifier()) {
-        result = result && getQualifier()
-            .equals(other.getQualifier());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasFamily()) {
-        hash = (37 * hash) + FAMILY_FIELD_NUMBER;
-        hash = (53 * hash) + getFamily().hashCode();
-      }
-      if (hasQualifier()) {
-        hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
-        hash = (53 * hash) + getQualifier().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code SumRequest}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        family_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        qualifier_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.family_ = family_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.qualifier_ = qualifier_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.getDefaultInstance()) return this;
-        if (other.hasFamily()) {
-          setFamily(other.getFamily());
-        }
-        if (other.hasQualifier()) {
-          setQualifier(other.getQualifier());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasFamily()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required bytes family = 1;
-      private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public boolean hasFamily() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public com.google.protobuf.ByteString getFamily() {
-        return family_;
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public Builder setFamily(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        family_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public Builder clearFamily() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        family_ = getDefaultInstance().getFamily();
-        onChanged();
-        return this;
-      }
-
-      // optional bytes qualifier = 2;
-      private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public boolean hasQualifier() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public com.google.protobuf.ByteString getQualifier() {
-        return qualifier_;
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public Builder setQualifier(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000002;
-        qualifier_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public Builder clearQualifier() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        qualifier_ = getDefaultInstance().getQualifier();
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:SumRequest)
-    }
-
-    static {
-      defaultInstance = new SumRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:SumRequest)
-  }
-
-  public interface SumResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required int64 sum = 1;
-    /**
-     * <code>required int64 sum = 1;</code>
-     */
-    boolean hasSum();
-    /**
-     * <code>required int64 sum = 1;</code>
-     */
-    long getSum();
-  }
-  /**
-   * Protobuf type {@code SumResponse}
-   */
-  public static final class SumResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements SumResponseOrBuilder {
-    // Use SumResponse.newBuilder() to construct.
-    private SumResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private SumResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final SumResponse defaultInstance;
-    public static SumResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public SumResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private SumResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              sum_ = input.readInt64();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<SumResponse> PARSER =
-        new com.google.protobuf.AbstractParser<SumResponse>() {
-      public SumResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new SumResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<SumResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required int64 sum = 1;
-    public static final int SUM_FIELD_NUMBER = 1;
-    private long sum_;
-    /**
-     * <code>required int64 sum = 1;</code>
-     */
-    public boolean hasSum() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required int64 sum = 1;</code>
-     */
-    public long getSum() {
-      return sum_;
-    }
-
-    private void initFields() {
-      sum_ = 0L;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasSum()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeInt64(1, sum_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt64Size(1, sum_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse) obj;
-
-      boolean result = true;
-      result = result && (hasSum() == other.hasSum());
-      if (hasSum()) {
-        result = result && (getSum()
-            == other.getSum());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasSum()) {
-        hash = (37 * hash) + SUM_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getSum());
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code SumResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        sum_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.internal_static_SumResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.sum_ = sum_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance()) return this;
-        if (other.hasSum()) {
-          setSum(other.getSum());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasSum()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required int64 sum = 1;
-      private long sum_ ;
-      /**
-       * <code>required int64 sum = 1;</code>
-       */
-      public boolean hasSum() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required int64 sum = 1;</code>
-       */
-      public long getSum() {
-        return sum_;
-      }
-      /**
-       * <code>required int64 sum = 1;</code>
-       */
-      public Builder setSum(long value) {
-        bitField0_ |= 0x00000001;
-        sum_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required int64 sum = 1;</code>
-       */
-      public Builder clearSum() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        sum_ = 0L;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:SumResponse)
-    }
-
-    static {
-      defaultInstance = new SumResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:SumResponse)
-  }
-
-  /**
-   * Protobuf service {@code ColumnAggregationService}
-   */
-  public static abstract class ColumnAggregationService
-      implements com.google.protobuf.Service {
-    protected ColumnAggregationService() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc sum(.SumRequest) returns (.SumResponse);</code>
-       */
-      public abstract void sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse> done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new ColumnAggregationService() {
-        @java.lang.Override
-        public  void sum(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse> done) {
-          impl.sum(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc sum(.SumRequest) returns (.SumResponse);</code>
-     */
-    public abstract void sum(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse> done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.ColumnAggregationService implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.class,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request)
-          throws com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationProtos.SumResponse.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:ColumnAggregationService)
-  }
-
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_SumRequest_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_SumRequest_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_SumResponse_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_SumResponse_fieldAccessorTable;
-
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\037ColumnAggregationProtocol.proto\"/\n\nSum" +
-      "Request\022\016\n\006family\030\001 \002(\014\022\021\n\tqualifier\030\002 \001" +
-      "(\014\"\032\n\013SumResponse\022\013\n\003sum\030\001 \002(\0032<\n\030Column" +
-      "AggregationService\022 \n\003sum\022\013.SumRequest\032\014" +
-      ".SumResponseBW\n6org.apache.hadoop.hbase." +
-      "coprocessor.protobuf.generatedB\027ColumnAg" +
-      "gregationProtos\210\001\001\240\001\001"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_SumRequest_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_SumRequest_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_SumRequest_descriptor,
-              new java.lang.String[] { "Family", "Qualifier", });
-          internal_static_SumResponse_descriptor =
-            getDescriptor().getMessageTypes().get(1);
-          internal_static_SumResponse_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_SumResponse_descriptor,
-              new java.lang.String[] { "Sum", });
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[35/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/CodedOutputStream.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/CodedOutputStream.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/CodedOutputStream.java
deleted file mode 100644
index 03871c9..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/CodedOutputStream.java
+++ /dev/null
@@ -1,3001 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.FIXED_32_SIZE;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.FIXED_64_SIZE;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.MAX_VARINT_SIZE;
-import static java.lang.Math.max;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Utf8.UnpairedSurrogateException;
-import java.io.IOException;
-import java.io.OutputStream;
-import java.nio.BufferOverflowException;
-import java.nio.ByteBuffer;
-import java.nio.ByteOrder;
-import java.util.logging.Level;
-import java.util.logging.Logger;
-
-/**
- * Encodes and writes protocol message fields.
- *
- * <p>This class contains two kinds of methods:  methods that write specific
- * protocol message constructs and field types (e.g. {@link #writeTag} and
- * {@link #writeInt32}) and methods that write low-level values (e.g.
- * {@link #writeRawVarint32} and {@link #writeRawBytes}).  If you are
- * writing encoded protocol messages, you should use the former methods, but if
- * you are writing some other format of your own design, use the latter.
- *
- * <p>This class is totally unsynchronized.
- */
-public abstract class CodedOutputStream extends ByteOutput {
-  private static final Logger logger = Logger.getLogger(CodedOutputStream.class.getName());
-  private static final boolean HAS_UNSAFE_ARRAY_OPERATIONS = UnsafeUtil.hasUnsafeArrayOperations();
-  private static final long ARRAY_BASE_OFFSET = UnsafeUtil.getArrayBaseOffset();
-
-  /**
-   * @deprecated Use {@link #computeFixed32SizeNoTag(int)} instead.
-   */
-  @Deprecated
-  public static final int LITTLE_ENDIAN_32_SIZE = FIXED_32_SIZE;
-
-  /**
-   * The buffer size used in {@link #newInstance(OutputStream)}.
-   */
-  public static final int DEFAULT_BUFFER_SIZE = 4096;
-
-  /**
-   * Returns the buffer size to efficiently write dataLength bytes to this
-   * CodedOutputStream. Used by AbstractMessageLite.
-   *
-   * @return the buffer size to efficiently write dataLength bytes to this
-   *         CodedOutputStream.
-   */
-  static int computePreferredBufferSize(int dataLength) {
-    if (dataLength > DEFAULT_BUFFER_SIZE) {
-      return DEFAULT_BUFFER_SIZE;
-    }
-    return dataLength;
-  }
-
-  /**
-   * Create a new {@code CodedOutputStream} wrapping the given {@code OutputStream}.
-   *
-   * <p> NOTE: The provided {@link OutputStream} <strong>MUST NOT</strong> retain access or
-   * modify the provided byte arrays. Doing so may result in corrupted data, which would be
-   * difficult to debug.
-   */
-  public static CodedOutputStream newInstance(final OutputStream output) {
-    return newInstance(output, DEFAULT_BUFFER_SIZE);
-  }
-
-  /**
-   * Create a new {@code CodedOutputStream} wrapping the given {@code OutputStream} with a given
-   * buffer size.
-   *
-   * <p> NOTE: The provided {@link OutputStream} <strong>MUST NOT</strong> retain access or
-   * modify the provided byte arrays. Doing so may result in corrupted data, which would be
-   * difficult to debug.
-   */
-  public static CodedOutputStream newInstance(final OutputStream output, final int bufferSize) {
-    return new OutputStreamEncoder(output, bufferSize);
-  }
-
-  /**
-   * Create a new {@code CodedOutputStream} that writes directly to the given
-   * byte array.  If more bytes are written than fit in the array,
-   * {@link OutOfSpaceException} will be thrown.  Writing directly to a flat
-   * array is faster than writing to an {@code OutputStream}.  See also
-   * {@link ByteString#newCodedBuilder}.
-   */
-  public static CodedOutputStream newInstance(final byte[] flatArray) {
-    return newInstance(flatArray, 0, flatArray.length);
-  }
-
-  /**
-   * Create a new {@code CodedOutputStream} that writes directly to the given
-   * byte array slice.  If more bytes are written than fit in the slice,
-   * {@link OutOfSpaceException} will be thrown.  Writing directly to a flat
-   * array is faster than writing to an {@code OutputStream}.  See also
-   * {@link ByteString#newCodedBuilder}.
-   */
-  public static CodedOutputStream newInstance(
-      final byte[] flatArray, final int offset, final int length) {
-    return new ArrayEncoder(flatArray, offset, length);
-  }
-
-  /** Create a new {@code CodedOutputStream} that writes to the given {@link ByteBuffer}. */
-  public static CodedOutputStream newInstance(ByteBuffer buffer) {
-    if (buffer.hasArray()) {
-      return new HeapNioEncoder(buffer);
-    }
-    if (buffer.isDirect() && !buffer.isReadOnly()) {
-      return UnsafeDirectNioEncoder.isSupported()
-          ? newUnsafeInstance(buffer)
-          : newSafeInstance(buffer);
-    }
-    throw new IllegalArgumentException("ByteBuffer is read-only");
-  }
-
-  /** For testing purposes only. */
-  static CodedOutputStream newUnsafeInstance(ByteBuffer buffer) {
-    return new UnsafeDirectNioEncoder(buffer);
-  }
-
-  /** For testing purposes only. */
-  static CodedOutputStream newSafeInstance(ByteBuffer buffer) {
-    return new SafeDirectNioEncoder(buffer);
-  }
-
-  /**
-   * Configures serialization to be deterministic.
-   *
-   * <p>The deterministic serialization guarantees that for a given binary, equal (defined by the
-   * {@code equals()} methods in protos) messages will always be serialized to the same bytes. This
-   * implies:
-   *
-   * <ul>
-   * <li>repeated serialization of a message will return the same bytes
-   * <li>different processes of the same binary (which may be executing on different machines) will
-   *     serialize equal messages to the same bytes.
-   * </ul>
-   *
-   * <p>Note the deterministic serialization is NOT canonical across languages; it is also unstable
-   * across different builds with schema changes due to unknown fields. Users who need canonical
-   * serialization, e.g. persistent storage in a canonical form, fingerprinting, etc, should define
-   * their own canonicalization specification and implement the serializer using reflection APIs
-   * rather than relying on this API.
-   *
-   * <p> Once set, the serializer will:  (Note this is an implementation detail and may subject to
-   * change in the future)
-   *
-   * <ul>
-   * <li> sort map entries by keys in lexicographical order or numerical order. Note: For string
-   *     keys, the order is based on comparing the Unicode value of each character in the strings.
-   *     The order may be different from the deterministic serialization in other languages where
-   *     maps are sorted on the lexicographical order of the UTF8 encoded keys.
-   * </ul>
-   */
-  void useDeterministicSerialization() {
-    serializationDeterministic = true;
-  }
-
-  boolean isSerializationDeterministic() {
-    return serializationDeterministic;
-  }
-  private boolean serializationDeterministic;
-
-  /**
-   * Create a new {@code CodedOutputStream} that writes to the given {@link ByteBuffer}.
-   *
-   * @deprecated the size parameter is no longer used since use of an internal buffer is useless
-   * (and wasteful) when writing to a {@link ByteBuffer}. Use {@link #newInstance(ByteBuffer)}
-   * instead.
-   */
-  @Deprecated
-  public static CodedOutputStream newInstance(ByteBuffer byteBuffer,
-      @SuppressWarnings("unused") int unused) {
-    return newInstance(byteBuffer);
-  }
-
-  /**
-   * Create a new {@code CodedOutputStream} that writes to the provided {@link ByteOutput}.
-   *
-   * <p> NOTE: The {@link ByteOutput} <strong>MUST NOT</strong> modify the provided buffers. Doing
-   * so may result in corrupted data, which would be difficult to debug.
-   *
-   * @param byteOutput the output target for encoded bytes.
-   * @param bufferSize the size of the internal scratch buffer to be used for string encoding.
-   * Setting this to {@code 0} will disable buffering, requiring an allocation for each encoded
-   * string.
-   */
-  static CodedOutputStream newInstance(ByteOutput byteOutput, int bufferSize) {
-    if (bufferSize < 0) {
-      throw new IllegalArgumentException("bufferSize must be positive");
-    }
-
-    return new ByteOutputEncoder(byteOutput, bufferSize);
-  }
-
-  // Disallow construction outside of this class.
-  private CodedOutputStream() {
-  }
-
-  // -----------------------------------------------------------------
-
-  /** Encode and write a tag. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeTag(int fieldNumber, int wireType) throws IOException;
-
-  /** Write an {@code int32} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeInt32(int fieldNumber, int value) throws IOException;
-
-  /** Write a {@code uint32} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeUInt32(int fieldNumber, int value) throws IOException;
-
-  /** Write a {@code sint32} field, including tag, to the stream. */
-  public final void writeSInt32(final int fieldNumber, final int value) throws IOException {
-    writeUInt32(fieldNumber, encodeZigZag32(value));
-  }
-
-  /** Write a {@code fixed32} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeFixed32(int fieldNumber, int value) throws IOException;
-
-  /** Write an {@code sfixed32} field, including tag, to the stream. */
-  public final void writeSFixed32(final int fieldNumber, final int value) throws IOException {
-    writeFixed32(fieldNumber, value);
-  }
-
-  /** Write an {@code int64} field, including tag, to the stream. */
-  public final void writeInt64(final int fieldNumber, final long value) throws IOException {
-    writeUInt64(fieldNumber, value);
-  }
-
-  /** Write a {@code uint64} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeUInt64(int fieldNumber, long value) throws IOException;
-
-  /** Write an {@code sint64} field, including tag, to the stream. */
-  public final void writeSInt64(final int fieldNumber, final long value) throws IOException {
-    writeUInt64(fieldNumber, encodeZigZag64(value));
-  }
-
-  /** Write a {@code fixed64} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeFixed64(int fieldNumber, long value) throws IOException;
-
-  /** Write an {@code sfixed64} field, including tag, to the stream. */
-  public final void writeSFixed64(final int fieldNumber, final long value) throws IOException {
-    writeFixed64(fieldNumber, value);
-  }
-
-  /** Write a {@code float} field, including tag, to the stream. */
-  public final void writeFloat(final int fieldNumber, final float value) throws IOException {
-    writeFixed32(fieldNumber, Float.floatToRawIntBits(value));
-  }
-
-  /** Write a {@code double} field, including tag, to the stream. */
-  public final void writeDouble(final int fieldNumber, final double value) throws IOException {
-    writeFixed64(fieldNumber, Double.doubleToRawLongBits(value));
-  }
-
-  /** Write a {@code bool} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeBool(int fieldNumber, boolean value) throws IOException;
-
-  /**
-   * Write an enum field, including tag, to the stream. The provided value is the numeric
-   * value used to represent the enum value on the wire (not the enum ordinal value).
-   */
-  public final void writeEnum(final int fieldNumber, final int value) throws IOException {
-    writeInt32(fieldNumber, value);
-  }
-
-  /** Write a {@code string} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeString(int fieldNumber, String value) throws IOException;
-
-  /** Write a {@code bytes} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeBytes(int fieldNumber, ByteString value) throws IOException;
-
-  /** Write a {@code bytes} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeByteArray(int fieldNumber, byte[] value) throws IOException;
-
-  /** Write a {@code bytes} field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeByteArray(int fieldNumber, byte[] value, int offset, int length)
-      throws IOException;
-
-  /**
-   * Write a {@code bytes} field, including tag, to the stream.
-   * This method will write all content of the ByteBuffer regardless of the
-   * current position and limit (i.e., the number of bytes to be written is
-   * value.capacity(), not value.remaining()). Furthermore, this method doesn't
-   * alter the state of the passed-in ByteBuffer. Its position, limit, mark,
-   * etc. will remain unchanged. If you only want to write the remaining bytes
-   * of a ByteBuffer, you can call
-   * {@code writeByteBuffer(fieldNumber, byteBuffer.slice())}.
-   */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeByteBuffer(int fieldNumber, ByteBuffer value) throws IOException;
-
-  /**
-   * Write a single byte.
-   */
-  public final void writeRawByte(final byte value) throws IOException {
-    write(value);
-  }
-
-  /** Write a single byte, represented by an integer value. */
-  public final void writeRawByte(final int value) throws IOException {
-    write((byte) value);
-  }
-
-  /** Write an array of bytes. */
-  public final void writeRawBytes(final byte[] value) throws IOException {
-    write(value, 0, value.length);
-  }
-
-  /**
-   * Write part of an array of bytes.
-   */
-  public final void writeRawBytes(final byte[] value, int offset, int length) throws IOException {
-    write(value, offset, length);
-  }
-
-  /** Write a byte string. */
-  public final void writeRawBytes(final ByteString value) throws IOException {
-    value.writeTo(this);
-  }
-
-  /**
-   * Write a ByteBuffer. This method will write all content of the ByteBuffer
-   * regardless of the current position and limit (i.e., the number of bytes
-   * to be written is value.capacity(), not value.remaining()). Furthermore,
-   * this method doesn't alter the state of the passed-in ByteBuffer. Its
-   * position, limit, mark, etc. will remain unchanged. If you only want to
-   * write the remaining bytes of a ByteBuffer, you can call
-   * {@code writeRawBytes(byteBuffer.slice())}.
-   */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeRawBytes(final ByteBuffer value) throws IOException;
-
-  /** Write an embedded message field, including tag, to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeMessage(final int fieldNumber, final MessageLite value)
-      throws IOException;
-
-  /**
-   * Write a MessageSet extension field to the stream.  For historical reasons,
-   * the wire format differs from normal fields.
-   */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeMessageSetExtension(final int fieldNumber, final MessageLite value)
-      throws IOException;
-
-  /**
-   * Write an unparsed MessageSet extension field to the stream.  For
-   * historical reasons, the wire format differs from normal fields.
-   */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeRawMessageSetExtension(final int fieldNumber, final ByteString value)
-      throws IOException;
-
-  // -----------------------------------------------------------------
-
-  /** Write an {@code int32} field to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeInt32NoTag(final int value) throws IOException;
-
-  /** Write a {@code uint32} field to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeUInt32NoTag(int value) throws IOException;
-
-  /** Write a {@code sint32} field to the stream. */
-  public final void writeSInt32NoTag(final int value) throws IOException {
-    writeUInt32NoTag(encodeZigZag32(value));
-  }
-
-  /** Write a {@code fixed32} field to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeFixed32NoTag(int value) throws IOException;
-
-  /** Write a {@code sfixed32} field to the stream. */
-  public final void writeSFixed32NoTag(final int value) throws IOException {
-    writeFixed32NoTag(value);
-  }
-
-  /** Write an {@code int64} field to the stream. */
-  public final void writeInt64NoTag(final long value) throws IOException {
-    writeUInt64NoTag(value);
-  }
-
-  /** Write a {@code uint64} field to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeUInt64NoTag(long value) throws IOException;
-
-  /** Write a {@code sint64} field to the stream. */
-  public final void writeSInt64NoTag(final long value) throws IOException {
-    writeUInt64NoTag(encodeZigZag64(value));
-  }
-
-  /** Write a {@code fixed64} field to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeFixed64NoTag(long value) throws IOException;
-
-  /** Write a {@code sfixed64} field to the stream. */
-  public final void writeSFixed64NoTag(final long value) throws IOException {
-    writeFixed64NoTag(value);
-  }
-
-  /** Write a {@code float} field to the stream. */
-  public final void writeFloatNoTag(final float value) throws IOException {
-    writeFixed32NoTag(Float.floatToRawIntBits(value));
-  }
-
-  /** Write a {@code double} field to the stream. */
-  public final void writeDoubleNoTag(final double value) throws IOException {
-    writeFixed64NoTag(Double.doubleToRawLongBits(value));
-  }
-
-  /** Write a {@code bool} field to the stream. */
-  public final void writeBoolNoTag(final boolean value) throws IOException {
-    write((byte) (value ? 1 : 0));
-  }
-
-  /**
-   * Write an enum field to the stream. The provided value is the numeric
-   * value used to represent the enum value on the wire (not the enum ordinal value).
-   */
-  public final void writeEnumNoTag(final int value) throws IOException {
-    writeInt32NoTag(value);
-  }
-
-  /** Write a {@code string} field to the stream. */
-  // TODO(dweis): Document behavior on ill-formed UTF-16 input.
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeStringNoTag(String value) throws IOException;
-
-  /** Write a {@code bytes} field to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeBytesNoTag(final ByteString value) throws IOException;
-
-  /** Write a {@code bytes} field to the stream. */
-  public final void writeByteArrayNoTag(final byte[] value) throws IOException {
-    writeByteArrayNoTag(value, 0, value.length);
-  }
-
-  /** Write an embedded message field to the stream. */
-  // Abstract to avoid overhead of additional virtual method calls.
-  public abstract void writeMessageNoTag(final MessageLite value) throws IOException;
-
-  //=================================================================
-
-  @ExperimentalApi
-  @Override
-  public abstract void write(byte value) throws IOException;
-
-  @ExperimentalApi
-  @Override
-  public abstract void write(byte[] value, int offset, int length) throws IOException;
-
-  @ExperimentalApi
-  @Override
-  public abstract void writeLazy(byte[] value, int offset, int length) throws IOException;
-
-  @Override
-  public abstract void write(ByteBuffer value) throws IOException;
-
-  @ExperimentalApi
-  @Override
-  public abstract void writeLazy(ByteBuffer value) throws IOException;
-
-  // =================================================================
-  // =================================================================
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code int32} field, including tag.
-   */
-  public static int computeInt32Size(final int fieldNumber, final int value) {
-    return computeTagSize(fieldNumber) + computeInt32SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code uint32} field, including tag.
-   */
-  public static int computeUInt32Size(final int fieldNumber, final int value) {
-    return computeTagSize(fieldNumber) + computeUInt32SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code sint32} field, including tag.
-   */
-  public static int computeSInt32Size(final int fieldNumber, final int value) {
-    return computeTagSize(fieldNumber) + computeSInt32SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code fixed32} field, including tag.
-   */
-  public static int computeFixed32Size(final int fieldNumber, final int value) {
-    return computeTagSize(fieldNumber) + computeFixed32SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code sfixed32} field, including tag.
-   */
-  public static int computeSFixed32Size(final int fieldNumber, final int value) {
-    return computeTagSize(fieldNumber) + computeSFixed32SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code int64} field, including tag.
-   */
-  public static int computeInt64Size(final int fieldNumber, final long value) {
-    return computeTagSize(fieldNumber) + computeInt64SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code uint64} field, including tag.
-   */
-  public static int computeUInt64Size(final int fieldNumber, final long value) {
-    return computeTagSize(fieldNumber) + computeUInt64SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code sint64} field, including tag.
-   */
-  public static int computeSInt64Size(final int fieldNumber, final long value) {
-    return computeTagSize(fieldNumber) + computeSInt64SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code fixed64} field, including tag.
-   */
-  public static int computeFixed64Size(final int fieldNumber, final long value) {
-    return computeTagSize(fieldNumber) + computeFixed64SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code sfixed64} field, including tag.
-   */
-  public static int computeSFixed64Size(final int fieldNumber, final long value) {
-    return computeTagSize(fieldNumber) + computeSFixed64SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code float} field, including tag.
-   */
-  public static int computeFloatSize(final int fieldNumber, final float value) {
-    return computeTagSize(fieldNumber) + computeFloatSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code double} field, including tag.
-   */
-  public static int computeDoubleSize(final int fieldNumber, final double value) {
-    return computeTagSize(fieldNumber) + computeDoubleSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code bool} field, including tag.
-   */
-  public static int computeBoolSize(final int fieldNumber, final boolean value) {
-    return computeTagSize(fieldNumber) + computeBoolSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * enum field, including tag. The provided value is the numeric
-   * value used to represent the enum value on the wire (not the enum ordinal value).
-   */
-  public static int computeEnumSize(final int fieldNumber, final int value) {
-    return computeTagSize(fieldNumber) + computeEnumSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code string} field, including tag.
-   */
-  public static int computeStringSize(final int fieldNumber, final String value) {
-    return computeTagSize(fieldNumber) + computeStringSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code bytes} field, including tag.
-   */
-  public static int computeBytesSize(final int fieldNumber, final ByteString value) {
-    return computeTagSize(fieldNumber) + computeBytesSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code bytes} field, including tag.
-   */
-  public static int computeByteArraySize(final int fieldNumber, final byte[] value) {
-    return computeTagSize(fieldNumber) + computeByteArraySizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code bytes} field, including tag.
-   */
-  public static int computeByteBufferSize(final int fieldNumber, final ByteBuffer value) {
-    return computeTagSize(fieldNumber) + computeByteBufferSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * embedded message in lazy field, including tag.
-   */
-  public static int computeLazyFieldSize(final int fieldNumber, final LazyFieldLite value) {
-    return computeTagSize(fieldNumber) + computeLazyFieldSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * embedded message field, including tag.
-   */
-  public static int computeMessageSize(final int fieldNumber, final MessageLite value) {
-    return computeTagSize(fieldNumber) + computeMessageSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * MessageSet extension to the stream.  For historical reasons,
-   * the wire format differs from normal fields.
-   */
-  public static int computeMessageSetExtensionSize(final int fieldNumber, final MessageLite value) {
-    return computeTagSize(WireFormat.MESSAGE_SET_ITEM) * 2
-        + computeUInt32Size(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber)
-        + computeMessageSize(WireFormat.MESSAGE_SET_MESSAGE, value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * unparsed MessageSet extension field to the stream.  For
-   * historical reasons, the wire format differs from normal fields.
-   */
-  public static int computeRawMessageSetExtensionSize(
-      final int fieldNumber, final ByteString value) {
-    return computeTagSize(WireFormat.MESSAGE_SET_ITEM) * 2
-        + computeUInt32Size(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber)
-        + computeBytesSize(WireFormat.MESSAGE_SET_MESSAGE, value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * lazily parsed MessageSet extension field to the stream.  For
-   * historical reasons, the wire format differs from normal fields.
-   */
-  public static int computeLazyFieldMessageSetExtensionSize(
-      final int fieldNumber, final LazyFieldLite value) {
-    return computeTagSize(WireFormat.MESSAGE_SET_ITEM) * 2
-        + computeUInt32Size(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber)
-        + computeLazyFieldSize(WireFormat.MESSAGE_SET_MESSAGE, value);
-  }
-
-  // -----------------------------------------------------------------
-
-  /** Compute the number of bytes that would be needed to encode a tag. */
-  public static int computeTagSize(final int fieldNumber) {
-    return computeUInt32SizeNoTag(WireFormat.makeTag(fieldNumber, 0));
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code int32} field, including tag.
-   */
-  public static int computeInt32SizeNoTag(final int value) {
-    if (value >= 0) {
-      return computeUInt32SizeNoTag(value);
-    } else {
-      // Must sign-extend.
-      return MAX_VARINT_SIZE;
-    }
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code uint32} field.
-   */
-  public static int computeUInt32SizeNoTag(final int value) {
-    if ((value & (~0 <<  7)) == 0) {
-      return 1;
-    }
-    if ((value & (~0 << 14)) == 0) {
-      return 2;
-    }
-    if ((value & (~0 << 21)) == 0) {
-      return 3;
-    }
-    if ((value & (~0 << 28)) == 0) {
-      return 4;
-    }
-    return 5;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code sint32} field.
-   */
-  public static int computeSInt32SizeNoTag(final int value) {
-    return computeUInt32SizeNoTag(encodeZigZag32(value));
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code fixed32} field.
-   */
-  public static int computeFixed32SizeNoTag(@SuppressWarnings("unused") final int unused) {
-    return FIXED_32_SIZE;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code sfixed32} field.
-   */
-  public static int computeSFixed32SizeNoTag(@SuppressWarnings("unused") final int unused) {
-    return FIXED_32_SIZE;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code int64} field, including tag.
-   */
-  public static int computeInt64SizeNoTag(final long value) {
-    return computeUInt64SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code uint64} field, including tag.
-   */
-  public static int computeUInt64SizeNoTag(long value) {
-    // handle two popular special cases up front ...
-    if ((value & (~0L << 7)) == 0L) {
-      return 1;
-    }
-    if (value < 0L) {
-      return 10;
-    }
-    // ... leaving us with 8 remaining, which we can divide and conquer
-    int n = 2;
-    if ((value & (~0L << 35)) != 0L) {
-      n += 4; value >>>= 28;
-    }
-    if ((value & (~0L << 21)) != 0L) {
-      n += 2; value >>>= 14;
-    }
-    if ((value & (~0L << 14)) != 0L) {
-      n += 1;
-    }
-    return n;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code sint64} field.
-   */
-  public static int computeSInt64SizeNoTag(final long value) {
-    return computeUInt64SizeNoTag(encodeZigZag64(value));
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code fixed64} field.
-   */
-  public static int computeFixed64SizeNoTag(@SuppressWarnings("unused") final long unused) {
-    return FIXED_64_SIZE;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an
-   * {@code sfixed64} field.
-   */
-  public static int computeSFixed64SizeNoTag(@SuppressWarnings("unused") final long unused) {
-    return FIXED_64_SIZE;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code float} field, including tag.
-   */
-  public static int computeFloatSizeNoTag(@SuppressWarnings("unused") final float unused) {
-    return FIXED_32_SIZE;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code double} field, including tag.
-   */
-  public static int computeDoubleSizeNoTag(@SuppressWarnings("unused") final double unused) {
-    return FIXED_64_SIZE;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code bool} field.
-   */
-  public static int computeBoolSizeNoTag(@SuppressWarnings("unused") final boolean unused) {
-    return 1;
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an enum field.
-   * The provided value is the numeric value used to represent the enum value on the wire
-   * (not the enum ordinal value).
-   */
-  public static int computeEnumSizeNoTag(final int value) {
-    return computeInt32SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code string} field.
-   */
-  public static int computeStringSizeNoTag(final String value) {
-    int length;
-    try {
-      length = Utf8.encodedLength(value);
-    } catch (UnpairedSurrogateException e) {
-      // TODO(dweis): Consider using nio Charset methods instead.
-      final byte[] bytes = value.getBytes(Internal.UTF_8);
-      length = bytes.length;
-    }
-
-    return computeLengthDelimitedFieldSize(length);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an embedded
-   * message stored in lazy field.
-   */
-  public static int computeLazyFieldSizeNoTag(final LazyFieldLite value) {
-    return computeLengthDelimitedFieldSize(value.getSerializedSize());
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code bytes} field.
-   */
-  public static int computeBytesSizeNoTag(final ByteString value) {
-    return computeLengthDelimitedFieldSize(value.size());
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code bytes} field.
-   */
-  public static int computeByteArraySizeNoTag(final byte[] value) {
-    return computeLengthDelimitedFieldSize(value.length);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code bytes} field.
-   */
-  public static int computeByteBufferSizeNoTag(final ByteBuffer value) {
-    return computeLengthDelimitedFieldSize(value.capacity());
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode an embedded
-   * message field.
-   */
-  public static int computeMessageSizeNoTag(final MessageLite value) {
-    return computeLengthDelimitedFieldSize(value.getSerializedSize());
-  }
-
-  static int computeLengthDelimitedFieldSize(int fieldLength) {
-    return computeUInt32SizeNoTag(fieldLength) + fieldLength;
-  }
-
-  /**
-   * Encode a ZigZag-encoded 32-bit value.  ZigZag encodes signed integers
-   * into values that can be efficiently encoded with varint.  (Otherwise,
-   * negative values must be sign-extended to 64 bits to be varint encoded,
-   * thus always taking 10 bytes on the wire.)
-   *
-   * @param n A signed 32-bit integer.
-   * @return An unsigned 32-bit integer, stored in a signed int because
-   *         Java has no explicit unsigned support.
-   */
-  public static int encodeZigZag32(final int n) {
-    // Note:  the right-shift must be arithmetic
-    return (n << 1) ^ (n >> 31);
-  }
-
-  /**
-   * Encode a ZigZag-encoded 64-bit value.  ZigZag encodes signed integers
-   * into values that can be efficiently encoded with varint.  (Otherwise,
-   * negative values must be sign-extended to 64 bits to be varint encoded,
-   * thus always taking 10 bytes on the wire.)
-   *
-   * @param n A signed 64-bit integer.
-   * @return An unsigned 64-bit integer, stored in a signed int because
-   *         Java has no explicit unsigned support.
-   */
-  public static long encodeZigZag64(final long n) {
-    // Note:  the right-shift must be arithmetic
-    return (n << 1) ^ (n >> 63);
-  }
-
-  // =================================================================
-
-  /**
-   * Flushes the stream and forces any buffered bytes to be written.  This
-   * does not flush the underlying OutputStream.
-   */
-  public abstract void flush() throws IOException;
-
-  /**
-   * If writing to a flat array, return the space left in the array.
-   * Otherwise, throws {@code UnsupportedOperationException}.
-   */
-  public abstract int spaceLeft();
-
-  /**
-   * Verifies that {@link #spaceLeft()} returns zero.  It's common to create
-   * a byte array that is exactly big enough to hold a message, then write to
-   * it with a {@code CodedOutputStream}.  Calling {@code checkNoSpaceLeft()}
-   * after writing verifies that the message was actually as big as expected,
-   * which can help catch bugs.
-   */
-  public final void checkNoSpaceLeft() {
-    if (spaceLeft() != 0) {
-      throw new IllegalStateException("Did not write as much data as expected.");
-    }
-  }
-
-  /**
-   * If you create a CodedOutputStream around a simple flat array, you must
-   * not attempt to write more bytes than the array has space.  Otherwise,
-   * this exception will be thrown.
-   */
-  public static class OutOfSpaceException extends IOException {
-    private static final long serialVersionUID = -6947486886997889499L;
-
-    private static final String MESSAGE =
-        "CodedOutputStream was writing to a flat byte array and ran out of space.";
-
-    OutOfSpaceException() {
-      super(MESSAGE);
-    }
-
-    OutOfSpaceException(String explanationMessage) {
-      super(MESSAGE + ": " + explanationMessage);
-    }
-
-    OutOfSpaceException(Throwable cause) {
-      super(MESSAGE, cause);
-    }
-
-    OutOfSpaceException(String explanationMessage, Throwable cause) {
-      super(MESSAGE + ": " + explanationMessage, cause);
-    }
-  }
-
-  /**
-   * Get the total number of bytes successfully written to this stream.  The
-   * returned value is not guaranteed to be accurate if exceptions have been
-   * found in the middle of writing.
-   */
-  public abstract int getTotalBytesWritten();
-
-  // =================================================================
-
-  /** Write a {@code bytes} field to the stream. Visible for testing. */
-  abstract void writeByteArrayNoTag(final byte[] value, final int offset, final int length)
-      throws IOException;
-
-  final void inefficientWriteStringNoTag(String value, UnpairedSurrogateException cause)
-      throws IOException {
-    logger.log(Level.WARNING,
-        "Converting ill-formed UTF-16. Your Protocol Buffer will not round trip correctly!", cause);
-
-    // Unfortunately there does not appear to be any way to tell Java to encode
-    // UTF-8 directly into our buffer, so we have to let it create its own byte
-    // array and then copy.
-    // TODO(dweis): Consider using nio Charset methods instead.
-    final byte[] bytes = value.getBytes(Internal.UTF_8);
-    try {
-      writeUInt32NoTag(bytes.length);
-      writeLazy(bytes, 0, bytes.length);
-    } catch (IndexOutOfBoundsException e) {
-      throw new OutOfSpaceException(e);
-    } catch (OutOfSpaceException e) {
-      throw e;
-    }
-  }
-
-  // =================================================================
-
-  /**
-   * Write a {@code group} field, including tag, to the stream.
-   *
-   * @deprecated groups are deprecated.
-   */
-  @Deprecated
-  public final void writeGroup(final int fieldNumber, final MessageLite value) throws IOException {
-    writeTag(fieldNumber, WireFormat.WIRETYPE_START_GROUP);
-    writeGroupNoTag(value);
-    writeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP);
-  }
-
-  /**
-   * Write a {@code group} field to the stream.
-   *
-   * @deprecated groups are deprecated.
-   */
-  @Deprecated
-  public final void writeGroupNoTag(final MessageLite value) throws IOException {
-    value.writeTo(this);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code group} field, including tag.
-   *
-   * @deprecated groups are deprecated.
-   */
-  @Deprecated
-  public static int computeGroupSize(final int fieldNumber, final MessageLite value) {
-    return computeTagSize(fieldNumber) * 2 + computeGroupSizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * {@code group} field.
-   */
-  @Deprecated
-  public static int computeGroupSizeNoTag(final MessageLite value) {
-    return value.getSerializedSize();
-  }
-
-  /**
-   * Encode and write a varint.  {@code value} is treated as
-   * unsigned, so it won't be sign-extended if negative.
-   *
-   * @deprecated use {@link #writeUInt32NoTag} instead.
-   */
-  @Deprecated
-  public final void writeRawVarint32(int value) throws IOException {
-    writeUInt32NoTag(value);
-  }
-
-  /**
-   * Encode and write a varint.
-   *
-   * @deprecated use {@link #writeUInt64NoTag} instead.
-   */
-  @Deprecated
-  public final void writeRawVarint64(long value) throws IOException {
-    writeUInt64NoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a varint.
-   * {@code value} is treated as unsigned, so it won't be sign-extended if
-   * negative.
-   *
-   * @deprecated use {@link #computeUInt32SizeNoTag(int)} instead.
-   */
-  @Deprecated
-  public static int computeRawVarint32Size(final int value) {
-    return computeUInt32SizeNoTag(value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a varint.
-   *
-   * @deprecated use {@link #computeUInt64SizeNoTag(long)} instead.
-   */
-  @Deprecated
-  public static int computeRawVarint64Size(long value) {
-    return computeUInt64SizeNoTag(value);
-  }
-
-  /**
-   * Write a little-endian 32-bit integer.
-   *
-   * @deprecated Use {@link #writeFixed32NoTag} instead.
-   */
-  @Deprecated
-  public final void writeRawLittleEndian32(final int value) throws IOException {
-    writeFixed32NoTag(value);
-  }
-
-  /**
-   * Write a little-endian 64-bit integer.
-   *
-   * @deprecated Use {@link #writeFixed64NoTag} instead.
-   */
-  @Deprecated
-  public final void writeRawLittleEndian64(final long value) throws IOException {
-    writeFixed64NoTag(value);
-  }
-
-  // =================================================================
-
-  /**
-   * A {@link CodedOutputStream} that writes directly to a byte array.
-   */
-  private static class ArrayEncoder extends CodedOutputStream {
-    private final byte[] buffer;
-    private final int offset;
-    private final int limit;
-    private int position;
-
-    ArrayEncoder(byte[] buffer, int offset, int length) {
-      if (buffer == null) {
-        throw new NullPointerException("buffer");
-      }
-      if ((offset | length | (buffer.length - (offset + length))) < 0) {
-        throw new IllegalArgumentException(String.format(
-            "Array range is invalid. Buffer.length=%d, offset=%d, length=%d",
-            buffer.length, offset, length));
-      }
-      this.buffer = buffer;
-      this.offset = offset;
-      position = offset;
-      limit = offset + length;
-    }
-
-    @Override
-    public final void writeTag(final int fieldNumber, final int wireType) throws IOException {
-      writeUInt32NoTag(WireFormat.makeTag(fieldNumber, wireType));
-    }
-
-    @Override
-    public final void writeInt32(final int fieldNumber, final int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeInt32NoTag(value);
-    }
-
-    @Override
-    public final void writeUInt32(final int fieldNumber, final int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeUInt32NoTag(value);
-    }
-
-    @Override
-    public final void writeFixed32(final int fieldNumber, final int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_FIXED32);
-      writeFixed32NoTag(value);
-    }
-
-    @Override
-    public final void writeUInt64(final int fieldNumber, final long value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeUInt64NoTag(value);
-    }
-
-    @Override
-    public final void writeFixed64(final int fieldNumber, final long value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_FIXED64);
-      writeFixed64NoTag(value);
-    }
-
-    @Override
-    public final void writeBool(final int fieldNumber, final boolean value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      write((byte) (value ? 1 : 0));
-    }
-
-    @Override
-    public final void writeString(final int fieldNumber, final String value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeStringNoTag(value);
-    }
-
-    @Override
-    public final void writeBytes(final int fieldNumber, final ByteString value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeBytesNoTag(value);
-    }
-
-    @Override
-    public final void writeByteArray(final int fieldNumber, final byte[] value) throws IOException {
-      writeByteArray(fieldNumber, value, 0, value.length);
-    }
-
-    @Override
-    public final void writeByteArray(
-        final int fieldNumber, final byte[] value, final int offset, final int length)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeByteArrayNoTag(value, offset, length);
-    }
-
-    @Override
-    public final void writeByteBuffer(final int fieldNumber, final ByteBuffer value)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeUInt32NoTag(value.capacity());
-      writeRawBytes(value);
-    }
-
-    @Override
-    public final void writeBytesNoTag(final ByteString value) throws IOException {
-      writeUInt32NoTag(value.size());
-      value.writeTo(this);
-    }
-
-    @Override
-    public final void writeByteArrayNoTag(final byte[] value, int offset, int length)
-        throws IOException {
-      writeUInt32NoTag(length);
-      write(value, offset, length);
-    }
-
-    @Override
-    public final void writeRawBytes(final ByteBuffer value) throws IOException {
-      if (value.hasArray()) {
-        write(value.array(), value.arrayOffset(), value.capacity());
-      } else {
-        ByteBuffer duplicated = value.duplicate();
-        duplicated.clear();
-        write(duplicated);
-      }
-    }
-
-    @Override
-    public final void writeMessage(final int fieldNumber, final MessageLite value)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeMessageNoTag(value);
-    }
-
-    @Override
-    public final void writeMessageSetExtension(final int fieldNumber, final MessageLite value)
-        throws IOException {
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_START_GROUP);
-      writeUInt32(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber);
-      writeMessage(WireFormat.MESSAGE_SET_MESSAGE, value);
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_END_GROUP);
-    }
-
-    @Override
-    public final void writeRawMessageSetExtension(final int fieldNumber, final ByteString value)
-        throws IOException {
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_START_GROUP);
-      writeUInt32(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber);
-      writeBytes(WireFormat.MESSAGE_SET_MESSAGE, value);
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_END_GROUP);
-    }
-
-    @Override
-    public final void writeMessageNoTag(final MessageLite value) throws IOException {
-      writeUInt32NoTag(value.getSerializedSize());
-      value.writeTo(this);
-    }
-
-    @Override
-    public final void write(byte value) throws IOException {
-      try {
-        buffer[position++] = value;
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, 1), e);
-      }
-    }
-
-    @Override
-    public final void writeInt32NoTag(int value) throws IOException {
-      if (value >= 0) {
-        writeUInt32NoTag(value);
-      } else {
-        // Must sign-extend.
-        writeUInt64NoTag(value);
-      }
-    }
-
-    @Override
-    public final void writeUInt32NoTag(int value) throws IOException {
-      if (HAS_UNSAFE_ARRAY_OPERATIONS && spaceLeft() >= MAX_VARINT_SIZE) {
-        long pos = ARRAY_BASE_OFFSET + position;
-        while (true) {
-          if ((value & ~0x7F) == 0) {
-            UnsafeUtil.putByte(buffer, pos++, (byte) value);
-            position++;
-            return;
-          } else {
-            UnsafeUtil.putByte(buffer, pos++, (byte) ((value & 0x7F) | 0x80));
-            position++;
-            value >>>= 7;
-          }
-        }
-      } else {
-        try {
-          while (true) {
-            if ((value & ~0x7F) == 0) {
-              buffer[position++] = (byte) value;
-              return;
-            } else {
-              buffer[position++] = (byte) ((value & 0x7F) | 0x80);
-              value >>>= 7;
-            }
-          }
-        } catch (IndexOutOfBoundsException e) {
-          throw new OutOfSpaceException(
-              String.format("Pos: %d, limit: %d, len: %d", position, limit, 1), e);
-        }
-      }
-    }
-
-    @Override
-    public final void writeFixed32NoTag(int value) throws IOException {
-      try {
-        buffer[position++] = (byte) (value & 0xFF);
-        buffer[position++] = (byte) ((value >> 8) & 0xFF);
-        buffer[position++] = (byte) ((value >> 16) & 0xFF);
-        buffer[position++] = (byte) ((value >> 24) & 0xFF);
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, 1), e);
-      }
-    }
-
-    @Override
-    public final void writeUInt64NoTag(long value) throws IOException {
-      if (HAS_UNSAFE_ARRAY_OPERATIONS && spaceLeft() >= MAX_VARINT_SIZE) {
-        long pos = ARRAY_BASE_OFFSET + position;
-        while (true) {
-          if ((value & ~0x7FL) == 0) {
-            UnsafeUtil.putByte(buffer, pos++, (byte) value);
-            position++;
-            return;
-          } else {
-            UnsafeUtil.putByte(buffer, pos++, (byte) (((int) value & 0x7F) | 0x80));
-            position++;
-            value >>>= 7;
-          }
-        }
-      } else {
-        try {
-          while (true) {
-            if ((value & ~0x7FL) == 0) {
-              buffer[position++] = (byte) value;
-              return;
-            } else {
-              buffer[position++] = (byte) (((int) value & 0x7F) | 0x80);
-              value >>>= 7;
-            }
-          }
-        } catch (IndexOutOfBoundsException e) {
-          throw new OutOfSpaceException(
-              String.format("Pos: %d, limit: %d, len: %d", position, limit, 1), e);
-        }
-      }
-    }
-
-    @Override
-    public final void writeFixed64NoTag(long value) throws IOException {
-      try {
-        buffer[position++] = (byte) ((int) (value) & 0xFF);
-        buffer[position++] = (byte) ((int) (value >> 8) & 0xFF);
-        buffer[position++] = (byte) ((int) (value >> 16) & 0xFF);
-        buffer[position++] = (byte) ((int) (value >> 24) & 0xFF);
-        buffer[position++] = (byte) ((int) (value >> 32) & 0xFF);
-        buffer[position++] = (byte) ((int) (value >> 40) & 0xFF);
-        buffer[position++] = (byte) ((int) (value >> 48) & 0xFF);
-        buffer[position++] = (byte) ((int) (value >> 56) & 0xFF);
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, 1), e);
-      }
-    }
-
-    @Override
-    public final void write(byte[] value, int offset, int length) throws IOException {
-      try {
-        System.arraycopy(value, offset, buffer, position, length);
-        position += length;
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, length), e);
-      }
-    }
-
-    @Override
-    public final void writeLazy(byte[] value, int offset, int length) throws IOException {
-      write(value, offset, length);
-    }
-
-    @Override
-    public final void write(ByteBuffer value) throws IOException {
-      final int length = value.remaining();
-      try {
-        value.get(buffer, position, length);
-        position += length;
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, length), e);
-      }
-    }
-
-    @Override
-    public final void writeLazy(ByteBuffer value) throws IOException {
-      write(value);
-    }
-
-    @Override
-    public final void writeStringNoTag(String value) throws IOException {
-      final int oldPosition = position;
-      try {
-        // UTF-8 byte length of the string is at least its UTF-16 code unit length (value.length()),
-        // and at most 3 times of it. We take advantage of this in both branches below.
-        final int maxLength = value.length() * Utf8.MAX_BYTES_PER_CHAR;
-        final int maxLengthVarIntSize = computeUInt32SizeNoTag(maxLength);
-        final int minLengthVarIntSize = computeUInt32SizeNoTag(value.length());
-        if (minLengthVarIntSize == maxLengthVarIntSize) {
-          position = oldPosition + minLengthVarIntSize;
-          int newPosition = Utf8.encode(value, buffer, position, spaceLeft());
-          // Since this class is stateful and tracks the position, we rewind and store the state,
-          // prepend the length, then reset it back to the end of the string.
-          position = oldPosition;
-          int length = newPosition - oldPosition - minLengthVarIntSize;
-          writeUInt32NoTag(length);
-          position = newPosition;
-        } else {
-          int length = Utf8.encodedLength(value);
-          writeUInt32NoTag(length);
-          position = Utf8.encode(value, buffer, position, spaceLeft());
-        }
-      } catch (UnpairedSurrogateException e) {
-        // Roll back the change - we fall back to inefficient path.
-        position = oldPosition;
-
-        // TODO(nathanmittler): We should throw an IOException here instead.
-        inefficientWriteStringNoTag(value, e);
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void flush() {
-      // Do nothing.
-    }
-
-    @Override
-    public final int spaceLeft() {
-      return limit - position;
-    }
-
-    @Override
-    public final int getTotalBytesWritten() {
-      return position - offset;
-    }
-  }
-
-  /**
-   * A {@link CodedOutputStream} that writes directly to a heap {@link ByteBuffer}. Writes are
-   * done directly to the underlying array. The buffer position is only updated after a flush.
-   */
-  private static final class HeapNioEncoder extends ArrayEncoder {
-    private final ByteBuffer byteBuffer;
-    private int initialPosition;
-
-    HeapNioEncoder(ByteBuffer byteBuffer) {
-      super(byteBuffer.array(), byteBuffer.arrayOffset() + byteBuffer.position(),
-          byteBuffer.remaining());
-      this.byteBuffer = byteBuffer;
-      this.initialPosition = byteBuffer.position();
-    }
-
-    @Override
-    public void flush() {
-      // Update the position on the buffer.
-      byteBuffer.position(initialPosition + getTotalBytesWritten());
-    }
-  }
-
-  /**
-   * A {@link CodedOutputStream} that writes directly to a direct {@link ByteBuffer}, using only
-   * safe operations..
-   */
-  private static final class SafeDirectNioEncoder extends CodedOutputStream {
-    private final ByteBuffer originalBuffer;
-    private final ByteBuffer buffer;
-    private final int initialPosition;
-
-    SafeDirectNioEncoder(ByteBuffer buffer) {
-      this.originalBuffer = buffer;
-      this.buffer = buffer.duplicate().order(ByteOrder.LITTLE_ENDIAN);
-      initialPosition = buffer.position();
-    }
-
-    @Override
-    public void writeTag(final int fieldNumber, final int wireType) throws IOException {
-      writeUInt32NoTag(WireFormat.makeTag(fieldNumber, wireType));
-    }
-
-    @Override
-    public void writeInt32(final int fieldNumber, final int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeInt32NoTag(value);
-    }
-
-    @Override
-    public void writeUInt32(final int fieldNumber, final int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeUInt32NoTag(value);
-    }
-
-    @Override
-    public void writeFixed32(final int fieldNumber, final int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_FIXED32);
-      writeFixed32NoTag(value);
-    }
-
-    @Override
-    public void writeUInt64(final int fieldNumber, final long value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeUInt64NoTag(value);
-    }
-
-    @Override
-    public void writeFixed64(final int fieldNumber, final long value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_FIXED64);
-      writeFixed64NoTag(value);
-    }
-
-    @Override
-    public void writeBool(final int fieldNumber, final boolean value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      write((byte) (value ? 1 : 0));
-    }
-
-    @Override
-    public void writeString(final int fieldNumber, final String value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeStringNoTag(value);
-    }
-
-    @Override
-    public void writeBytes(final int fieldNumber, final ByteString value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeBytesNoTag(value);
-    }
-
-    @Override
-    public void writeByteArray(final int fieldNumber, final byte[] value) throws IOException {
-      writeByteArray(fieldNumber, value, 0, value.length);
-    }
-
-    @Override
-    public void writeByteArray(
-        final int fieldNumber, final byte[] value, final int offset, final int length)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeByteArrayNoTag(value, offset, length);
-    }
-
-    @Override
-    public void writeByteBuffer(final int fieldNumber, final ByteBuffer value)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeUInt32NoTag(value.capacity());
-      writeRawBytes(value);
-    }
-
-    @Override
-    public void writeMessage(final int fieldNumber, final MessageLite value)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeMessageNoTag(value);
-    }
-
-    @Override
-    public void writeMessageSetExtension(final int fieldNumber, final MessageLite value)
-        throws IOException {
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_START_GROUP);
-      writeUInt32(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber);
-      writeMessage(WireFormat.MESSAGE_SET_MESSAGE, value);
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_END_GROUP);
-    }
-
-    @Override
-    public void writeRawMessageSetExtension(final int fieldNumber, final ByteString value)
-        throws IOException {
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_START_GROUP);
-      writeUInt32(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber);
-      writeBytes(WireFormat.MESSAGE_SET_MESSAGE, value);
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_END_GROUP);
-    }
-
-    @Override
-    public void writeMessageNoTag(final MessageLite value) throws IOException {
-      writeUInt32NoTag(value.getSerializedSize());
-      value.writeTo(this);
-    }
-
-    @Override
-    public void write(byte value) throws IOException {
-      try {
-        buffer.put(value);
-      } catch (BufferOverflowException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void writeBytesNoTag(final ByteString value) throws IOException {
-      writeUInt32NoTag(value.size());
-      value.writeTo(this);
-    }
-
-    @Override
-    public void writeByteArrayNoTag(final byte[] value, int offset, int length) throws IOException {
-      writeUInt32NoTag(length);
-      write(value, offset, length);
-    }
-
-    @Override
-    public void writeRawBytes(final ByteBuffer value) throws IOException {
-      if (value.hasArray()) {
-        write(value.array(), value.arrayOffset(), value.capacity());
-      } else {
-        ByteBuffer duplicated = value.duplicate();
-        duplicated.clear();
-        write(duplicated);
-      }
-    }
-
-    @Override
-    public void writeInt32NoTag(int value) throws IOException {
-      if (value >= 0) {
-        writeUInt32NoTag(value);
-      } else {
-        // Must sign-extend.
-        writeUInt64NoTag(value);
-      }
-    }
-
-    @Override
-    public void writeUInt32NoTag(int value) throws IOException {
-      try {
-        while (true) {
-          if ((value & ~0x7F) == 0) {
-            buffer.put((byte) value);
-            return;
-          } else {
-            buffer.put((byte) ((value & 0x7F) | 0x80));
-            value >>>= 7;
-          }
-        }
-      } catch (BufferOverflowException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void writeFixed32NoTag(int value) throws IOException {
-      try {
-        buffer.putInt(value);
-      } catch (BufferOverflowException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void writeUInt64NoTag(long value) throws IOException {
-      try {
-        while (true) {
-          if ((value & ~0x7FL) == 0) {
-            buffer.put((byte) value);
-            return;
-          } else {
-            buffer.put((byte) (((int) value & 0x7F) | 0x80));
-            value >>>= 7;
-          }
-        }
-      } catch (BufferOverflowException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void writeFixed64NoTag(long value) throws IOException {
-      try {
-        buffer.putLong(value);
-      } catch (BufferOverflowException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void write(byte[] value, int offset, int length) throws IOException {
-      try {
-        buffer.put(value, offset, length);
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(e);
-      } catch (BufferOverflowException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void writeLazy(byte[] value, int offset, int length) throws IOException {
-      write(value, offset, length);
-    }
-
-    @Override
-    public void write(ByteBuffer value) throws IOException {
-      try {
-        buffer.put(value);
-      } catch (BufferOverflowException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void writeLazy(ByteBuffer value) throws IOException {
-      write(value);
-    }
-
-    @Override
-    public void writeStringNoTag(String value) throws IOException {
-      final int startPos = buffer.position();
-      try {
-        // UTF-8 byte length of the string is at least its UTF-16 code unit length (value.length()),
-        // and at most 3 times of it. We take advantage of this in both branches below.
-        final int maxEncodedSize = value.length() * Utf8.MAX_BYTES_PER_CHAR;
-        final int maxLengthVarIntSize = computeUInt32SizeNoTag(maxEncodedSize);
-        final int minLengthVarIntSize = computeUInt32SizeNoTag(value.length());
-        if (minLengthVarIntSize == maxLengthVarIntSize) {
-          // Save the current position and increment past the length field. We'll come back
-          // and write the length field after the encoding is complete.
-          final int startOfBytes = buffer.position() + minLengthVarIntSize;
-          buffer.position(startOfBytes);
-
-          // Encode the string.
-          encode(value);
-
-          // Now go back to the beginning and write the length.
-          int endOfBytes = buffer.position();
-          buffer.position(startPos);
-          writeUInt32NoTag(endOfBytes - startOfBytes);
-
-          // Reposition the buffer past the written data.
-          buffer.position(endOfBytes);
-        } else {
-          final int length = Utf8.encodedLength(value);
-          writeUInt32NoTag(length);
-          encode(value);
-        }
-      } catch (UnpairedSurrogateException e) {
-        // Roll back the change and convert to an IOException.
-        buffer.position(startPos);
-
-        // TODO(nathanmittler): We should throw an IOException here instead.
-        inefficientWriteStringNoTag(value, e);
-      } catch (IllegalArgumentException e) {
-        // Thrown by buffer.position() if out of range.
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void flush() {
-      // Update the position of the original buffer.
-      originalBuffer.position(buffer.position());
-    }
-
-    @Override
-    public int spaceLeft() {
-      return buffer.remaining();
-    }
-
-    @Override
-    public int getTotalBytesWritten() {
-      return buffer.position() - initialPosition;
-    }
-
-    private void encode(String value) throws IOException {
-      try {
-        Utf8.encodeUtf8(value, buffer);
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-  }
-
-  /**
-   * A {@link CodedOutputStream} that writes directly to a direct {@link ByteBuffer} using {@code
-   * sun.misc.Unsafe}.
-   */
-  private static final class UnsafeDirectNioEncoder extends CodedOutputStream {
-    private final ByteBuffer originalBuffer;
-    private final ByteBuffer buffer;
-    private final long address;
-    private final long initialPosition;
-    private final long limit;
-    private final long oneVarintLimit;
-    private long position;
-
-    UnsafeDirectNioEncoder(ByteBuffer buffer) {
-      this.originalBuffer = buffer;
-      this.buffer = buffer.duplicate().order(ByteOrder.LITTLE_ENDIAN);
-      address = UnsafeUtil.addressOffset(buffer);
-      initialPosition = address + buffer.position();
-      limit = address + buffer.limit();
-      oneVarintLimit = limit - MAX_VARINT_SIZE;
-      position = initialPosition;
-    }
-
-    static boolean isSupported() {
-      return UnsafeUtil.hasUnsafeByteBufferOperations();
-    }
-
-    @Override
-    public void writeTag(int fieldNumber, int wireType) throws IOException {
-      writeUInt32NoTag(WireFormat.makeTag(fieldNumber, wireType));
-    }
-
-    @Override
-    public void writeInt32(int fieldNumber, int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeInt32NoTag(value);
-    }
-
-    @Override
-    public void writeUInt32(int fieldNumber, int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeUInt32NoTag(value);
-    }
-
-    @Override
-    public void writeFixed32(int fieldNumber, int value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_FIXED32);
-      writeFixed32NoTag(value);
-    }
-
-    @Override
-    public void writeUInt64(int fieldNumber, long value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      writeUInt64NoTag(value);
-    }
-
-    @Override
-    public void writeFixed64(int fieldNumber, long value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_FIXED64);
-      writeFixed64NoTag(value);
-    }
-
-    @Override
-    public void writeBool(int fieldNumber, boolean value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      write((byte) (value ? 1 : 0));
-    }
-
-    @Override
-    public void writeString(int fieldNumber, String value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeStringNoTag(value);
-    }
-
-    @Override
-    public void writeBytes(int fieldNumber, ByteString value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeBytesNoTag(value);
-    }
-
-    @Override
-    public void writeByteArray(int fieldNumber, byte[] value) throws IOException {
-      writeByteArray(fieldNumber, value, 0, value.length);
-    }
-
-    @Override
-    public void writeByteArray(int fieldNumber, byte[] value, int offset, int length)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeByteArrayNoTag(value, offset, length);
-    }
-
-    @Override
-    public void writeByteBuffer(int fieldNumber, ByteBuffer value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeUInt32NoTag(value.capacity());
-      writeRawBytes(value);
-    }
-
-    @Override
-    public void writeMessage(int fieldNumber, MessageLite value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeMessageNoTag(value);
-    }
-
-    @Override
-    public void writeMessageSetExtension(int fieldNumber, MessageLite value) throws IOException {
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_START_GROUP);
-      writeUInt32(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber);
-      writeMessage(WireFormat.MESSAGE_SET_MESSAGE, value);
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_END_GROUP);
-    }
-
-    @Override
-    public void writeRawMessageSetExtension(int fieldNumber, ByteString value) throws IOException {
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_START_GROUP);
-      writeUInt32(WireFormat.MESSAGE_SET_TYPE_ID, fieldNumber);
-      writeBytes(WireFormat.MESSAGE_SET_MESSAGE, value);
-      writeTag(WireFormat.MESSAGE_SET_ITEM, WireFormat.WIRETYPE_END_GROUP);
-    }
-
-    @Override
-    public void writeMessageNoTag(MessageLite value) throws IOException {
-      writeUInt32NoTag(value.getSerializedSize());
-      value.writeTo(this);
-    }
-
-    @Override
-    public void write(byte value) throws IOException {
-      if (position >= limit) {
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, 1));
-      }
-      UnsafeUtil.putByte(position++, value);
-    }
-
-    @Override
-    public void writeBytesNoTag(ByteString value) throws IOException {
-      writeUInt32NoTag(value.size());
-      value.writeTo(this);
-    }
-
-    @Override
-    public void writeByteArrayNoTag(byte[] value, int offset, int length) throws IOException {
-      writeUInt32NoTag(length);
-      write(value, offset, length);
-    }
-
-    @Override
-    public void writeRawBytes(ByteBuffer value) throws IOException {
-      if (value.hasArray()) {
-        write(value.array(), value.arrayOffset(), value.capacity());
-      } else {
-        ByteBuffer duplicated = value.duplicate();
-        duplicated.clear();
-        write(duplicated);
-      }
-    }
-
-    @Override
-    public void writeInt32NoTag(int value) throws IOException {
-      if (value >= 0) {
-        writeUInt32NoTag(value);
-      } else {
-        // Must sign-extend.
-        writeUInt64NoTag(value);
-      }
-    }
-
-    @Override
-    public void writeUInt32NoTag(int value) throws IOException {
-      if (position <= oneVarintLimit) {
-        // Optimization to avoid bounds checks on each iteration.
-        while (true) {
-          if ((value & ~0x7F) == 0) {
-            UnsafeUtil.putByte(position++, (byte) value);
-            return;
-          } else {
-            UnsafeUtil.putByte(position++, (byte) ((value & 0x7F) | 0x80));
-            value >>>= 7;
-          }
-        }
-      } else {
-        while (position < limit) {
-          if ((value & ~0x7F) == 0) {
-            UnsafeUtil.putByte(position++, (byte) value);
-            return;
-          } else {
-            UnsafeUtil.putByte(position++, (byte) ((value & 0x7F) | 0x80));
-            value >>>= 7;
-          }
-        }
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, 1));
-      }
-    }
-
-    @Override
-    public void writeFixed32NoTag(int value) throws IOException {
-      buffer.putInt(bufferPos(position), value);
-      position += FIXED_32_SIZE;
-    }
-
-    @Override
-    public void writeUInt64NoTag(long value) throws IOException {
-      if (position <= oneVarintLimit) {
-        // Optimization to avoid bounds checks on each iteration.
-        while (true) {
-          if ((value & ~0x7FL) == 0) {
-            UnsafeUtil.putByte(position++, (byte) value);
-            return;
-          } else {
-            UnsafeUtil.putByte(position++, (byte) (((int) value & 0x7F) | 0x80));
-            value >>>= 7;
-          }
-        }
-      } else {
-        while (position < limit) {
-          if ((value & ~0x7FL) == 0) {
-            UnsafeUtil.putByte(position++, (byte) value);
-            return;
-          } else {
-            UnsafeUtil.putByte(position++, (byte) (((int) value & 0x7F) | 0x80));
-            value >>>= 7;
-          }
-        }
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, 1));
-      }
-    }
-
-    @Override
-    public void writeFixed64NoTag(long value) throws IOException {
-      buffer.putLong(bufferPos(position), value);
-      position += FIXED_64_SIZE;
-    }
-
-    @Override
-    public void write(byte[] value, int offset, int length) throws IOException {
-      if (value == null
-          || offset < 0
-          || length < 0
-          || (value.length - length) < offset
-          || (limit - length) < position) {
-        if (value == null) {
-          throw new NullPointerException("value");
-        }
-        throw new OutOfSpaceException(
-            String.format("Pos: %d, limit: %d, len: %d", position, limit, length));
-      }
-
-      UnsafeUtil.copyMemory(
-          value, UnsafeUtil.getArrayBaseOffset() + offset, null, position, length);
-      position += length;
-    }
-
-    @Override
-    public void writeLazy(byte[] value, int offset, int length) throws IOException {
-      write(value, offset, length);
-    }
-
-    @Override
-    public void write(ByteBuffer value) throws IOException {
-      try {
-        int length = value.remaining();
-        repositionBuffer(position);
-        buffer.put(value);
-        position += length;
-      } catch (BufferOverflowException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void writeLazy(ByteBuffer value) throws IOException {
-      write(value);
-    }
-
-    @Override
-    public void writeStringNoTag(String value) throws IOException {
-      long prevPos = position;
-      try {
-        // UTF-8 byte length of the string is at least its UTF-16 code unit length (value.length()),
-        // and at most 3 times of it. We take advantage of this in both branches below.
-        int maxEncodedSize = value.length() * Utf8.MAX_BYTES_PER_CHAR;
-        int maxLengthVarIntSize = computeUInt32SizeNoTag(maxEncodedSize);
-        int minLengthVarIntSize = computeUInt32SizeNoTag(value.length());
-        if (minLengthVarIntSize == maxLengthVarIntSize) {
-          // Save the current position and increment past the length field. We'll come back
-          // and write the length field after the encoding is complete.
-          int stringStart = bufferPos(position) + minLengthVarIntSize;
-          buffer.position(stringStart);
-
-          // Encode the string.
-          Utf8.encodeUtf8(value, buffer);
-
-          // Write the length and advance the position.
-          int length = buffer.position() - stringStart;
-          writeUInt32NoTag(length);
-          position += length;
-        } else {
-          // Calculate and write the encoded length.
-          int length = Utf8.encodedLength(value);
-          writeUInt32NoTag(length);
-
-          // Write the string and advance the position.
-          repositionBuffer(position);
-          Utf8.encodeUtf8(value, buffer);
-          position += length;
-        }
-      } catch (UnpairedSurrogateException e) {
-        // Roll back the change and convert to an IOException.
-        position = prevPos;
-        repositionBuffer(position);
-
-        // TODO(nathanmittler): We should throw an IOException here instead.
-        inefficientWriteStringNoTag(value, e);
-      } catch (IllegalArgumentException e) {
-        // Thrown by buffer.position() if out of range.
-        throw new OutOfSpaceException(e);
-      } catch (IndexOutOfBoundsException e) {
-        throw new OutOfSpaceException(e);
-      }
-    }
-
-    @Override
-    public void flush() {
-      // Update the position of the original buffer.
-      originalBuffer.position(bufferPos(position));
-    }
-
-    @Override
-    public int spaceLeft() {
-      return (int) (limit - position);
-    }
-
-    @Override
-    public int getTotalBytesWritten() {
-      return (int) (position - initialPosition);
-    }
-
-    private void repositionBuffer(long pos) {
-      buffer.position(bufferPos(pos));
-    }
-
-    private int bufferPos(long pos) {
-      return (int) (pos - address);
-    }
-  }
-
-  /**
-   * Abstract base class for buffered encoders.
-   */
-  private abstract static class AbstractBufferedEncoder extends CodedOutputStream {
-    final byte[] buffer;
-    final int limit;
-    int position;
-    int totalBytesWritten;
-
-    AbstractBufferedEncoder(int bufferSize) {
-      if (bufferSize < 0) {
-        throw new IllegalArgumentException("bufferSize must be >= 0");
-      }
-      // As an optimization, we require that the buffer be able to store at least 2
-      // varints so that we can buffer any integer write (tag + value). This reduces the
-      // number of range checks for a single write to 1 (i.e. if there is not enough space
-      // to buffer the tag+value, flush and then buffer it).
-      this.buffer = new byte[max(bufferSize, MAX_VARINT_SIZE * 2)];
-      this.limit = buffer.length;
-    }
-
-    @Override
-    public final int spaceLeft() {
-      throw new UnsupportedOperationException(
-          "spaceLeft() can only be called on CodedOutputStreams that are "
-              + "writing to a flat array or ByteBuffer.");
-    }
-
-    @Override
-    public final int getTotalBytesWritten() {
-      return totalBytesWritten;
-    }
-
-    /**
-     * This method does not perform bounds checking on the array. Checking array bounds is the
-     * responsibility of the caller.
-     */
-    final void buffer(byte value) {
-      buffer[position++] = value;
-      totalBytesWritten++;
-    }
-
-    /**
-     * This method does not perform bounds checking on the array. Checking array bounds is the
-     * responsibility of the caller.
-     */
-    final void bufferTag(final int fieldNumber, final int wireType) {
-      bufferUInt32NoTag(WireFormat.makeTag(fieldNumber, wireType));
-    }
-
-    /**
-     * This method does not perform bounds checking on the array. Checking array bounds is the
-     * responsibility of the caller.
-     */
-    final void bufferInt32NoTag(final int value) {
-      if (value >= 0) {
-        bufferUInt32NoTag(value);
-      } else {
-        // Must sign-extend.
-        bufferUInt64NoTag(value);
-      }
-    }
-
-    /**
-     * This method does not perform bounds checking on the array. Checking array bounds is the
-     * responsibility of the caller.
-     */
-    final void bufferUInt32NoTag(int value) {
-      if (HAS_UNSAFE_ARRAY_OPERATIONS) {
-        final long originalPos = ARRAY_BASE_OFFSET + position;
-        long pos = originalPos;
-        while (true) {
-          if ((value & ~0x7F) == 0) {
-            UnsafeUtil.putByte(buffer, pos++, (byte) value);
-            break;
-          } else {
-            UnsafeUtil.putByte(buffer, pos++, (byte) ((value & 0x7F) | 0x80));
-            value >>>= 7;
-          }
-        }
-        int delta = (int) (pos - originalPos);
-        position += delta;
-        totalBytesWritten += delta;
-      } else {
-        while (true) {
-          if ((value & ~0x7F) == 0) {
-            buffer[position++] = (byte) value;
-            totalBytesWritten++;
-            return;
-          } else {
-            buffer[position++] = (byte) ((value & 0x7F) | 0x80);
-            totalBytesWritten++;
-            value >>>= 7;
-          }
-        }
-      }
-    }
-
-    /**
-     * This method does not perform bounds checking on the array. Checking array bounds is the
-     * responsibility of the caller.
-     */
-    final void bufferUInt64NoTag(long value) {
-      if (HAS_UNSAFE_ARRAY_OPERATIONS) {
-        final long originalPos = ARRAY_BASE_OFFSET + position;
-        long pos = originalPos;
-        while (true) {
-          if ((value & ~0x7FL) == 0) {
-            UnsafeUtil.putByte(buffer, pos++, (byte) value);
-            break;
-          } else {
-            UnsafeUtil.putByte(buffer, pos++, (byte) (((int) value & 0x7F) | 0x80));
-            value >>>= 7;
-          }
-        }
-        int delta = (int) (pos - originalPos);
-        position += delta;
-        totalBytesWritten += delta;
-      } else {
-        while (true) {
-          if ((value & ~0x7FL) == 0) {
-            buffer[position++] = (byte) value;
-            totalBytesWritten++;
-            return;
-          } else {
-            buffer[position++] = (byte) (((int) value & 0x7F) | 0x80);
-            totalBytesWritten++;
-            value >>>= 7;
-          }
-        }
-      }
-    }
-
-    /**
-     * This method does not perform bounds checking on the array. Checking array bounds is the
-     * responsibility of the caller.
-     */
-    final void bufferFixed32NoTag(int value) {
-      buffer[position++] = (byte) (value & 0xFF);
-      buffer[position++] = (byte) ((value >> 8) & 0xFF);
-      buffer[position++] = (byte) ((value >> 16) & 0xFF);
-      buffer[position++] = (byte) ((value >> 24) & 0xFF);
-      totalBytesWritten += FIXED_32_SIZE;
-    }
-
-    /**
-     * This method does not perform bounds checking on the array. Checking array bounds is the
-     * responsibility of the caller.
-     */
-    final void bufferFixed64NoTag(long value) {
-      buffer[position++] = (byte) (value & 0xFF);
-      buffer[position++] = (byte) ((value >> 8) & 0xFF);
-      buffer[position++] = (byte) ((value >> 16) & 0xFF);
-      buffer[position++] = (byte) ((value >> 24) & 0xFF);
-      buffer[position++] = (byte) ((int) (value >> 32) & 0xFF);
-      buffer[position++] = (byte) ((int) (value >> 40) & 0xFF);
-      buffer[position++] = (byte) ((int) (value >> 48) & 0xFF);
-      buffer[position++] = (byte) ((int) (value >> 56) & 0xFF);
-      totalBytesWritten += FIXED_64_SIZE;
-    }
-  }
-
-  /**
-   * A {@link CodedOutputStream} that decorates a {@link ByteOutput}. It internal buffer only to
-   * support string encoding operations. All other writes are just passed through to the
-   * {@link ByteOutput}.
-   */
-  private static final class ByteOutputEncoder extends AbstractBufferedEncoder {
-    private final ByteOutput out;
-
-    ByteOutputEncoder(ByteOutput out, int bufferSize) {
-      super(bufferSize);
-      if (out == null) {
-        throw new NullPointerException("out");
-      }
-      this.out = out;
-    }
-
-    @Override
-    public void writeTag(final int fieldNumber, final int wireType) throws IOException {
-      writeUInt32NoTag(WireFormat.makeTag(fieldNumber, wireType));
-    }
-
-    @Override
-    public void writeInt32(final int fieldNumber, final int value) throws IOException {
-      flushIfNotAvailable(MAX_VARINT_SIZE * 2);
-      bufferTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      bufferInt32NoTag(value);
-    }
-
-    @Override
-    public void writeUInt32(final int fieldNumber, final int value) throws IOException {
-      flushIfNotAvailable(MAX_VARINT_SIZE * 2);
-      bufferTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      bufferUInt32NoTag(value);
-    }
-
-    @Override
-    public void writeFixed32(final int fieldNumber, final int value) throws IOException {
-      flushIfNotAvailable(MAX_VARINT_SIZE + FIXED_32_SIZE);
-      bufferTag(fieldNumber, WireFormat.WIRETYPE_FIXED32);
-      bufferFixed32NoTag(value);
-    }
-
-    @Override
-    public void writeUInt64(final int fieldNumber, final long value) throws IOException {
-      flushIfNotAvailable(MAX_VARINT_SIZE * 2);
-      bufferTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      bufferUInt64NoTag(value);
-    }
-
-    @Override
-    public void writeFixed64(final int fieldNumber, final long value) throws IOException {
-      flushIfNotAvailable(MAX_VARINT_SIZE + FIXED_64_SIZE);
-      bufferTag(fieldNumber, WireFormat.WIRETYPE_FIXED64);
-      bufferFixed64NoTag(value);
-    }
-
-    @Override
-    public void writeBool(final int fieldNumber, final boolean value) throws IOException {
-      flushIfNotAvailable(MAX_VARINT_SIZE + 1);
-      bufferTag(fieldNumber, WireFormat.WIRETYPE_VARINT);
-      buffer((byte) (value ? 1 : 0));
-    }
-
-    @Override
-    public void writeString(final int fieldNumber, final String value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeStringNoTag(value);
-    }
-
-    @Override
-    public void writeBytes(final int fieldNumber, final ByteString value) throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeBytesNoTag(value);
-    }
-
-    @Override
-    public void writeByteArray(final int fieldNumber, final byte[] value) throws IOException {
-      writeByteArray(fieldNumber, value, 0, value.length);
-    }
-
-    @Override
-    public void writeByteArray(
-        final int fieldNumber, final byte[] value, final int offset, final int length)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeByteArrayNoTag(value, offset, length);
-    }
-
-    @Override
-    public void writeByteBuffer(final int fieldNumber, final ByteBuffer value)
-        throws IOException {
-      writeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      writeUInt32NoTag(value.capacity());
-      writeRawBytes(value);
-    }
-
-    @Override
-    public void writeBytesNoTag(final ByteString value) throws IOException {
-      writeUInt32NoTag(value.size());
-      value.writeTo(this);
-    }
-
-    @Override
-    public void writeByteArrayNoTag(final byte[] value, int offset, int length) throws IOException {
-      writeUInt32NoTag(length);
-      write(value, offset, length);
-    }
-
-    @Override
-    public void writeRawBytes(final ByteBuffer value) throws IOException {
-      if (value.hasArray()) {
-        write(value.array(), value.arrayOffset(), value.capacity());
-      } else {
-        ByteBuffer duplicated = value.duplicate();
-        duplicated.clear();
-        write(duplicated);

<TRUNCATED>

[44/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java
----------------------------------------------------------------------
diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java
deleted file mode 100644
index 508790c..0000000
--- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/protobuf/generated/AggregateProtos.java
+++ /dev/null
@@ -1,2375 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: Aggregate.proto
-
-package org.apache.hadoop.hbase.protobuf.generated;
-
-public final class AggregateProtos {
-  private AggregateProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface AggregateRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required string interpreter_class_name = 1;
-    /**
-     * <code>required string interpreter_class_name = 1;</code>
-     *
-     * <pre>
-     ** The request passed to the AggregateService consists of three parts
-     *  (1) the (canonical) classname of the ColumnInterpreter implementation
-     *  (2) the Scan query
-     *  (3) any bytes required to construct the ColumnInterpreter object
-     *      properly
-     * </pre>
-     */
-    boolean hasInterpreterClassName();
-    /**
-     * <code>required string interpreter_class_name = 1;</code>
-     *
-     * <pre>
-     ** The request passed to the AggregateService consists of three parts
-     *  (1) the (canonical) classname of the ColumnInterpreter implementation
-     *  (2) the Scan query
-     *  (3) any bytes required to construct the ColumnInterpreter object
-     *      properly
-     * </pre>
-     */
-    java.lang.String getInterpreterClassName();
-    /**
-     * <code>required string interpreter_class_name = 1;</code>
-     *
-     * <pre>
-     ** The request passed to the AggregateService consists of three parts
-     *  (1) the (canonical) classname of the ColumnInterpreter implementation
-     *  (2) the Scan query
-     *  (3) any bytes required to construct the ColumnInterpreter object
-     *      properly
-     * </pre>
-     */
-    com.google.protobuf.ByteString
-        getInterpreterClassNameBytes();
-
-    // required .hbase.pb.Scan scan = 2;
-    /**
-     * <code>required .hbase.pb.Scan scan = 2;</code>
-     */
-    boolean hasScan();
-    /**
-     * <code>required .hbase.pb.Scan scan = 2;</code>
-     */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan();
-    /**
-     * <code>required .hbase.pb.Scan scan = 2;</code>
-     */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder();
-
-    // optional bytes interpreter_specific_bytes = 3;
-    /**
-     * <code>optional bytes interpreter_specific_bytes = 3;</code>
-     */
-    boolean hasInterpreterSpecificBytes();
-    /**
-     * <code>optional bytes interpreter_specific_bytes = 3;</code>
-     */
-    com.google.protobuf.ByteString getInterpreterSpecificBytes();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.AggregateRequest}
-   */
-  public static final class AggregateRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements AggregateRequestOrBuilder {
-    // Use AggregateRequest.newBuilder() to construct.
-    private AggregateRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private AggregateRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final AggregateRequest defaultInstance;
-    public static AggregateRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public AggregateRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private AggregateRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              interpreterClassName_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000002) == 0x00000002)) {
-                subBuilder = scan_.toBuilder();
-              }
-              scan_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(scan_);
-                scan_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000002;
-              break;
-            }
-            case 26: {
-              bitField0_ |= 0x00000004;
-              interpreterSpecificBytes_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.class, org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<AggregateRequest> PARSER =
-        new com.google.protobuf.AbstractParser<AggregateRequest>() {
-      public AggregateRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new AggregateRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<AggregateRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required string interpreter_class_name = 1;
-    public static final int INTERPRETER_CLASS_NAME_FIELD_NUMBER = 1;
-    private java.lang.Object interpreterClassName_;
-    /**
-     * <code>required string interpreter_class_name = 1;</code>
-     *
-     * <pre>
-     ** The request passed to the AggregateService consists of three parts
-     *  (1) the (canonical) classname of the ColumnInterpreter implementation
-     *  (2) the Scan query
-     *  (3) any bytes required to construct the ColumnInterpreter object
-     *      properly
-     * </pre>
-     */
-    public boolean hasInterpreterClassName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required string interpreter_class_name = 1;</code>
-     *
-     * <pre>
-     ** The request passed to the AggregateService consists of three parts
-     *  (1) the (canonical) classname of the ColumnInterpreter implementation
-     *  (2) the Scan query
-     *  (3) any bytes required to construct the ColumnInterpreter object
-     *      properly
-     * </pre>
-     */
-    public java.lang.String getInterpreterClassName() {
-      java.lang.Object ref = interpreterClassName_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        com.google.protobuf.ByteString bs = 
-            (com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          interpreterClassName_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <code>required string interpreter_class_name = 1;</code>
-     *
-     * <pre>
-     ** The request passed to the AggregateService consists of three parts
-     *  (1) the (canonical) classname of the ColumnInterpreter implementation
-     *  (2) the Scan query
-     *  (3) any bytes required to construct the ColumnInterpreter object
-     *      properly
-     * </pre>
-     */
-    public com.google.protobuf.ByteString
-        getInterpreterClassNameBytes() {
-      java.lang.Object ref = interpreterClassName_;
-      if (ref instanceof java.lang.String) {
-        com.google.protobuf.ByteString b = 
-            com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        interpreterClassName_ = b;
-        return b;
-      } else {
-        return (com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    // required .hbase.pb.Scan scan = 2;
-    public static final int SCAN_FIELD_NUMBER = 2;
-    private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan scan_;
-    /**
-     * <code>required .hbase.pb.Scan scan = 2;</code>
-     */
-    public boolean hasScan() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>required .hbase.pb.Scan scan = 2;</code>
-     */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan() {
-      return scan_;
-    }
-    /**
-     * <code>required .hbase.pb.Scan scan = 2;</code>
-     */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder() {
-      return scan_;
-    }
-
-    // optional bytes interpreter_specific_bytes = 3;
-    public static final int INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER = 3;
-    private com.google.protobuf.ByteString interpreterSpecificBytes_;
-    /**
-     * <code>optional bytes interpreter_specific_bytes = 3;</code>
-     */
-    public boolean hasInterpreterSpecificBytes() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>optional bytes interpreter_specific_bytes = 3;</code>
-     */
-    public com.google.protobuf.ByteString getInterpreterSpecificBytes() {
-      return interpreterSpecificBytes_;
-    }
-
-    private void initFields() {
-      interpreterClassName_ = "";
-      scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance();
-      interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasInterpreterClassName()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!hasScan()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!getScan().isInitialized()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, getInterpreterClassNameBytes());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeMessage(2, scan_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeBytes(3, interpreterSpecificBytes_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, getInterpreterClassNameBytes());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(2, scan_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(3, interpreterSpecificBytes_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest other = (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest) obj;
-
-      boolean result = true;
-      result = result && (hasInterpreterClassName() == other.hasInterpreterClassName());
-      if (hasInterpreterClassName()) {
-        result = result && getInterpreterClassName()
-            .equals(other.getInterpreterClassName());
-      }
-      result = result && (hasScan() == other.hasScan());
-      if (hasScan()) {
-        result = result && getScan()
-            .equals(other.getScan());
-      }
-      result = result && (hasInterpreterSpecificBytes() == other.hasInterpreterSpecificBytes());
-      if (hasInterpreterSpecificBytes()) {
-        result = result && getInterpreterSpecificBytes()
-            .equals(other.getInterpreterSpecificBytes());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasInterpreterClassName()) {
-        hash = (37 * hash) + INTERPRETER_CLASS_NAME_FIELD_NUMBER;
-        hash = (53 * hash) + getInterpreterClassName().hashCode();
-      }
-      if (hasScan()) {
-        hash = (37 * hash) + SCAN_FIELD_NUMBER;
-        hash = (53 * hash) + getScan().hashCode();
-      }
-      if (hasInterpreterSpecificBytes()) {
-        hash = (37 * hash) + INTERPRETER_SPECIFIC_BYTES_FIELD_NUMBER;
-        hash = (53 * hash) + getInterpreterSpecificBytes().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.AggregateRequest}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.class, org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-          getScanFieldBuilder();
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        interpreterClassName_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
-        if (scanBuilder_ == null) {
-          scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance();
-        } else {
-          scanBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000002);
-        interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000004);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest build() {
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest buildPartial() {
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest result = new org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.interpreterClassName_ = interpreterClassName_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        if (scanBuilder_ == null) {
-          result.scan_ = scan_;
-        } else {
-          result.scan_ = scanBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        result.interpreterSpecificBytes_ = interpreterSpecificBytes_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest other) {
-        if (other == org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance()) return this;
-        if (other.hasInterpreterClassName()) {
-          bitField0_ |= 0x00000001;
-          interpreterClassName_ = other.interpreterClassName_;
-          onChanged();
-        }
-        if (other.hasScan()) {
-          mergeScan(other.getScan());
-        }
-        if (other.hasInterpreterSpecificBytes()) {
-          setInterpreterSpecificBytes(other.getInterpreterSpecificBytes());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasInterpreterClassName()) {
-          
-          return false;
-        }
-        if (!hasScan()) {
-          
-          return false;
-        }
-        if (!getScan().isInitialized()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required string interpreter_class_name = 1;
-      private java.lang.Object interpreterClassName_ = "";
-      /**
-       * <code>required string interpreter_class_name = 1;</code>
-       *
-       * <pre>
-       ** The request passed to the AggregateService consists of three parts
-       *  (1) the (canonical) classname of the ColumnInterpreter implementation
-       *  (2) the Scan query
-       *  (3) any bytes required to construct the ColumnInterpreter object
-       *      properly
-       * </pre>
-       */
-      public boolean hasInterpreterClassName() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required string interpreter_class_name = 1;</code>
-       *
-       * <pre>
-       ** The request passed to the AggregateService consists of three parts
-       *  (1) the (canonical) classname of the ColumnInterpreter implementation
-       *  (2) the Scan query
-       *  (3) any bytes required to construct the ColumnInterpreter object
-       *      properly
-       * </pre>
-       */
-      public java.lang.String getInterpreterClassName() {
-        java.lang.Object ref = interpreterClassName_;
-        if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          interpreterClassName_ = s;
-          return s;
-        } else {
-          return (java.lang.String) ref;
-        }
-      }
-      /**
-       * <code>required string interpreter_class_name = 1;</code>
-       *
-       * <pre>
-       ** The request passed to the AggregateService consists of three parts
-       *  (1) the (canonical) classname of the ColumnInterpreter implementation
-       *  (2) the Scan query
-       *  (3) any bytes required to construct the ColumnInterpreter object
-       *      properly
-       * </pre>
-       */
-      public com.google.protobuf.ByteString
-          getInterpreterClassNameBytes() {
-        java.lang.Object ref = interpreterClassName_;
-        if (ref instanceof String) {
-          com.google.protobuf.ByteString b = 
-              com.google.protobuf.ByteString.copyFromUtf8(
-                  (java.lang.String) ref);
-          interpreterClassName_ = b;
-          return b;
-        } else {
-          return (com.google.protobuf.ByteString) ref;
-        }
-      }
-      /**
-       * <code>required string interpreter_class_name = 1;</code>
-       *
-       * <pre>
-       ** The request passed to the AggregateService consists of three parts
-       *  (1) the (canonical) classname of the ColumnInterpreter implementation
-       *  (2) the Scan query
-       *  (3) any bytes required to construct the ColumnInterpreter object
-       *      properly
-       * </pre>
-       */
-      public Builder setInterpreterClassName(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        interpreterClassName_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string interpreter_class_name = 1;</code>
-       *
-       * <pre>
-       ** The request passed to the AggregateService consists of three parts
-       *  (1) the (canonical) classname of the ColumnInterpreter implementation
-       *  (2) the Scan query
-       *  (3) any bytes required to construct the ColumnInterpreter object
-       *      properly
-       * </pre>
-       */
-      public Builder clearInterpreterClassName() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        interpreterClassName_ = getDefaultInstance().getInterpreterClassName();
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string interpreter_class_name = 1;</code>
-       *
-       * <pre>
-       ** The request passed to the AggregateService consists of three parts
-       *  (1) the (canonical) classname of the ColumnInterpreter implementation
-       *  (2) the Scan query
-       *  (3) any bytes required to construct the ColumnInterpreter object
-       *      properly
-       * </pre>
-       */
-      public Builder setInterpreterClassNameBytes(
-          com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        interpreterClassName_ = value;
-        onChanged();
-        return this;
-      }
-
-      // required .hbase.pb.Scan scan = 2;
-      private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder> scanBuilder_;
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      public boolean hasScan() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan() {
-        if (scanBuilder_ == null) {
-          return scan_;
-        } else {
-          return scanBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      public Builder setScan(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan value) {
-        if (scanBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          scan_ = value;
-          onChanged();
-        } else {
-          scanBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000002;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      public Builder setScan(
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder builderForValue) {
-        if (scanBuilder_ == null) {
-          scan_ = builderForValue.build();
-          onChanged();
-        } else {
-          scanBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000002;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      public Builder mergeScan(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan value) {
-        if (scanBuilder_ == null) {
-          if (((bitField0_ & 0x00000002) == 0x00000002) &&
-              scan_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance()) {
-            scan_ =
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.newBuilder(scan_).mergeFrom(value).buildPartial();
-          } else {
-            scan_ = value;
-          }
-          onChanged();
-        } else {
-          scanBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000002;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      public Builder clearScan() {
-        if (scanBuilder_ == null) {
-          scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance();
-          onChanged();
-        } else {
-          scanBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder getScanBuilder() {
-        bitField0_ |= 0x00000002;
-        onChanged();
-        return getScanFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder() {
-        if (scanBuilder_ != null) {
-          return scanBuilder_.getMessageOrBuilder();
-        } else {
-          return scan_;
-        }
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 2;</code>
-       */
-      private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder> 
-          getScanFieldBuilder() {
-        if (scanBuilder_ == null) {
-          scanBuilder_ = new com.google.protobuf.SingleFieldBuilder<
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder>(
-                  scan_,
-                  getParentForChildren(),
-                  isClean());
-          scan_ = null;
-        }
-        return scanBuilder_;
-      }
-
-      // optional bytes interpreter_specific_bytes = 3;
-      private com.google.protobuf.ByteString interpreterSpecificBytes_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>optional bytes interpreter_specific_bytes = 3;</code>
-       */
-      public boolean hasInterpreterSpecificBytes() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
-      }
-      /**
-       * <code>optional bytes interpreter_specific_bytes = 3;</code>
-       */
-      public com.google.protobuf.ByteString getInterpreterSpecificBytes() {
-        return interpreterSpecificBytes_;
-      }
-      /**
-       * <code>optional bytes interpreter_specific_bytes = 3;</code>
-       */
-      public Builder setInterpreterSpecificBytes(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000004;
-        interpreterSpecificBytes_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bytes interpreter_specific_bytes = 3;</code>
-       */
-      public Builder clearInterpreterSpecificBytes() {
-        bitField0_ = (bitField0_ & ~0x00000004);
-        interpreterSpecificBytes_ = getDefaultInstance().getInterpreterSpecificBytes();
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.AggregateRequest)
-    }
-
-    static {
-      defaultInstance = new AggregateRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.AggregateRequest)
-  }
-
-  public interface AggregateResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // repeated bytes first_part = 1;
-    /**
-     * <code>repeated bytes first_part = 1;</code>
-     *
-     * <pre>
-     **
-     * The AggregateService methods all have a response that either is a Pair
-     * or a simple object. When it is a Pair both first_part and second_part
-     * have defined values (and the second_part is not present in the response
-     * when the response is not a pair). Refer to the AggregateImplementation 
-     * class for an overview of the AggregateResponse object constructions. 
-     * </pre>
-     */
-    java.util.List<com.google.protobuf.ByteString> getFirstPartList();
-    /**
-     * <code>repeated bytes first_part = 1;</code>
-     *
-     * <pre>
-     **
-     * The AggregateService methods all have a response that either is a Pair
-     * or a simple object. When it is a Pair both first_part and second_part
-     * have defined values (and the second_part is not present in the response
-     * when the response is not a pair). Refer to the AggregateImplementation 
-     * class for an overview of the AggregateResponse object constructions. 
-     * </pre>
-     */
-    int getFirstPartCount();
-    /**
-     * <code>repeated bytes first_part = 1;</code>
-     *
-     * <pre>
-     **
-     * The AggregateService methods all have a response that either is a Pair
-     * or a simple object. When it is a Pair both first_part and second_part
-     * have defined values (and the second_part is not present in the response
-     * when the response is not a pair). Refer to the AggregateImplementation 
-     * class for an overview of the AggregateResponse object constructions. 
-     * </pre>
-     */
-    com.google.protobuf.ByteString getFirstPart(int index);
-
-    // optional bytes second_part = 2;
-    /**
-     * <code>optional bytes second_part = 2;</code>
-     */
-    boolean hasSecondPart();
-    /**
-     * <code>optional bytes second_part = 2;</code>
-     */
-    com.google.protobuf.ByteString getSecondPart();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.AggregateResponse}
-   */
-  public static final class AggregateResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements AggregateResponseOrBuilder {
-    // Use AggregateResponse.newBuilder() to construct.
-    private AggregateResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private AggregateResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final AggregateResponse defaultInstance;
-    public static AggregateResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public AggregateResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private AggregateResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-                firstPart_ = new java.util.ArrayList<com.google.protobuf.ByteString>();
-                mutable_bitField0_ |= 0x00000001;
-              }
-              firstPart_.add(input.readBytes());
-              break;
-            }
-            case 18: {
-              bitField0_ |= 0x00000001;
-              secondPart_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-          firstPart_ = java.util.Collections.unmodifiableList(firstPart_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.class, org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<AggregateResponse> PARSER =
-        new com.google.protobuf.AbstractParser<AggregateResponse>() {
-      public AggregateResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new AggregateResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<AggregateResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // repeated bytes first_part = 1;
-    public static final int FIRST_PART_FIELD_NUMBER = 1;
-    private java.util.List<com.google.protobuf.ByteString> firstPart_;
-    /**
-     * <code>repeated bytes first_part = 1;</code>
-     *
-     * <pre>
-     **
-     * The AggregateService methods all have a response that either is a Pair
-     * or a simple object. When it is a Pair both first_part and second_part
-     * have defined values (and the second_part is not present in the response
-     * when the response is not a pair). Refer to the AggregateImplementation 
-     * class for an overview of the AggregateResponse object constructions. 
-     * </pre>
-     */
-    public java.util.List<com.google.protobuf.ByteString>
-        getFirstPartList() {
-      return firstPart_;
-    }
-    /**
-     * <code>repeated bytes first_part = 1;</code>
-     *
-     * <pre>
-     **
-     * The AggregateService methods all have a response that either is a Pair
-     * or a simple object. When it is a Pair both first_part and second_part
-     * have defined values (and the second_part is not present in the response
-     * when the response is not a pair). Refer to the AggregateImplementation 
-     * class for an overview of the AggregateResponse object constructions. 
-     * </pre>
-     */
-    public int getFirstPartCount() {
-      return firstPart_.size();
-    }
-    /**
-     * <code>repeated bytes first_part = 1;</code>
-     *
-     * <pre>
-     **
-     * The AggregateService methods all have a response that either is a Pair
-     * or a simple object. When it is a Pair both first_part and second_part
-     * have defined values (and the second_part is not present in the response
-     * when the response is not a pair). Refer to the AggregateImplementation 
-     * class for an overview of the AggregateResponse object constructions. 
-     * </pre>
-     */
-    public com.google.protobuf.ByteString getFirstPart(int index) {
-      return firstPart_.get(index);
-    }
-
-    // optional bytes second_part = 2;
-    public static final int SECOND_PART_FIELD_NUMBER = 2;
-    private com.google.protobuf.ByteString secondPart_;
-    /**
-     * <code>optional bytes second_part = 2;</code>
-     */
-    public boolean hasSecondPart() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>optional bytes second_part = 2;</code>
-     */
-    public com.google.protobuf.ByteString getSecondPart() {
-      return secondPart_;
-    }
-
-    private void initFields() {
-      firstPart_ = java.util.Collections.emptyList();
-      secondPart_ = com.google.protobuf.ByteString.EMPTY;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      for (int i = 0; i < firstPart_.size(); i++) {
-        output.writeBytes(1, firstPart_.get(i));
-      }
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(2, secondPart_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      {
-        int dataSize = 0;
-        for (int i = 0; i < firstPart_.size(); i++) {
-          dataSize += com.google.protobuf.CodedOutputStream
-            .computeBytesSizeNoTag(firstPart_.get(i));
-        }
-        size += dataSize;
-        size += 1 * getFirstPartList().size();
-      }
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(2, secondPart_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse other = (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse) obj;
-
-      boolean result = true;
-      result = result && getFirstPartList()
-          .equals(other.getFirstPartList());
-      result = result && (hasSecondPart() == other.hasSecondPart());
-      if (hasSecondPart()) {
-        result = result && getSecondPart()
-            .equals(other.getSecondPart());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (getFirstPartCount() > 0) {
-        hash = (37 * hash) + FIRST_PART_FIELD_NUMBER;
-        hash = (53 * hash) + getFirstPartList().hashCode();
-      }
-      if (hasSecondPart()) {
-        hash = (37 * hash) + SECOND_PART_FIELD_NUMBER;
-        hash = (53 * hash) + getSecondPart().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.AggregateResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.class, org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        firstPart_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000001);
-        secondPart_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.internal_static_hbase_pb_AggregateResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse build() {
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse buildPartial() {
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse result = new org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((bitField0_ & 0x00000001) == 0x00000001)) {
-          firstPart_ = java.util.Collections.unmodifiableList(firstPart_);
-          bitField0_ = (bitField0_ & ~0x00000001);
-        }
-        result.firstPart_ = firstPart_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.secondPart_ = secondPart_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse other) {
-        if (other == org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance()) return this;
-        if (!other.firstPart_.isEmpty()) {
-          if (firstPart_.isEmpty()) {
-            firstPart_ = other.firstPart_;
-            bitField0_ = (bitField0_ & ~0x00000001);
-          } else {
-            ensureFirstPartIsMutable();
-            firstPart_.addAll(other.firstPart_);
-          }
-          onChanged();
-        }
-        if (other.hasSecondPart()) {
-          setSecondPart(other.getSecondPart());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // repeated bytes first_part = 1;
-      private java.util.List<com.google.protobuf.ByteString> firstPart_ = java.util.Collections.emptyList();
-      private void ensureFirstPartIsMutable() {
-        if (!((bitField0_ & 0x00000001) == 0x00000001)) {
-          firstPart_ = new java.util.ArrayList<com.google.protobuf.ByteString>(firstPart_);
-          bitField0_ |= 0x00000001;
-         }
-      }
-      /**
-       * <code>repeated bytes first_part = 1;</code>
-       *
-       * <pre>
-       **
-       * The AggregateService methods all have a response that either is a Pair
-       * or a simple object. When it is a Pair both first_part and second_part
-       * have defined values (and the second_part is not present in the response
-       * when the response is not a pair). Refer to the AggregateImplementation 
-       * class for an overview of the AggregateResponse object constructions. 
-       * </pre>
-       */
-      public java.util.List<com.google.protobuf.ByteString>
-          getFirstPartList() {
-        return java.util.Collections.unmodifiableList(firstPart_);
-      }
-      /**
-       * <code>repeated bytes first_part = 1;</code>
-       *
-       * <pre>
-       **
-       * The AggregateService methods all have a response that either is a Pair
-       * or a simple object. When it is a Pair both first_part and second_part
-       * have defined values (and the second_part is not present in the response
-       * when the response is not a pair). Refer to the AggregateImplementation 
-       * class for an overview of the AggregateResponse object constructions. 
-       * </pre>
-       */
-      public int getFirstPartCount() {
-        return firstPart_.size();
-      }
-      /**
-       * <code>repeated bytes first_part = 1;</code>
-       *
-       * <pre>
-       **
-       * The AggregateService methods all have a response that either is a Pair
-       * or a simple object. When it is a Pair both first_part and second_part
-       * have defined values (and the second_part is not present in the response
-       * when the response is not a pair). Refer to the AggregateImplementation 
-       * class for an overview of the AggregateResponse object constructions. 
-       * </pre>
-       */
-      public com.google.protobuf.ByteString getFirstPart(int index) {
-        return firstPart_.get(index);
-      }
-      /**
-       * <code>repeated bytes first_part = 1;</code>
-       *
-       * <pre>
-       **
-       * The AggregateService methods all have a response that either is a Pair
-       * or a simple object. When it is a Pair both first_part and second_part
-       * have defined values (and the second_part is not present in the response
-       * when the response is not a pair). Refer to the AggregateImplementation 
-       * class for an overview of the AggregateResponse object constructions. 
-       * </pre>
-       */
-      public Builder setFirstPart(
-          int index, com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureFirstPartIsMutable();
-        firstPart_.set(index, value);
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>repeated bytes first_part = 1;</code>
-       *
-       * <pre>
-       **
-       * The AggregateService methods all have a response that either is a Pair
-       * or a simple object. When it is a Pair both first_part and second_part
-       * have defined values (and the second_part is not present in the response
-       * when the response is not a pair). Refer to the AggregateImplementation 
-       * class for an overview of the AggregateResponse object constructions. 
-       * </pre>
-       */
-      public Builder addFirstPart(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureFirstPartIsMutable();
-        firstPart_.add(value);
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>repeated bytes first_part = 1;</code>
-       *
-       * <pre>
-       **
-       * The AggregateService methods all have a response that either is a Pair
-       * or a simple object. When it is a Pair both first_part and second_part
-       * have defined values (and the second_part is not present in the response
-       * when the response is not a pair). Refer to the AggregateImplementation 
-       * class for an overview of the AggregateResponse object constructions. 
-       * </pre>
-       */
-      public Builder addAllFirstPart(
-          java.lang.Iterable<? extends com.google.protobuf.ByteString> values) {
-        ensureFirstPartIsMutable();
-        super.addAll(values, firstPart_);
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>repeated bytes first_part = 1;</code>
-       *
-       * <pre>
-       **
-       * The AggregateService methods all have a response that either is a Pair
-       * or a simple object. When it is a Pair both first_part and second_part
-       * have defined values (and the second_part is not present in the response
-       * when the response is not a pair). Refer to the AggregateImplementation 
-       * class for an overview of the AggregateResponse object constructions. 
-       * </pre>
-       */
-      public Builder clearFirstPart() {
-        firstPart_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000001);
-        onChanged();
-        return this;
-      }
-
-      // optional bytes second_part = 2;
-      private com.google.protobuf.ByteString secondPart_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>optional bytes second_part = 2;</code>
-       */
-      public boolean hasSecondPart() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional bytes second_part = 2;</code>
-       */
-      public com.google.protobuf.ByteString getSecondPart() {
-        return secondPart_;
-      }
-      /**
-       * <code>optional bytes second_part = 2;</code>
-       */
-      public Builder setSecondPart(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000002;
-        secondPart_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bytes second_part = 2;</code>
-       */
-      public Builder clearSecondPart() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        secondPart_ = getDefaultInstance().getSecondPart();
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.AggregateResponse)
-    }
-
-    static {
-      defaultInstance = new AggregateResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.AggregateResponse)
-  }
-
-  /**
-   * Protobuf service {@code hbase.pb.AggregateService}
-   *
-   * <pre>
-   ** Refer to the AggregateImplementation class for an overview of the 
-   *  AggregateService method implementations and their functionality.
-   * </pre>
-   */
-  public static abstract class AggregateService
-      implements com.google.protobuf.Service {
-    protected AggregateService() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc GetMax(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-       */
-      public abstract void getMax(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-      /**
-       * <code>rpc GetMin(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-       */
-      public abstract void getMin(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-      /**
-       * <code>rpc GetSum(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-       */
-      public abstract void getSum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-      /**
-       * <code>rpc GetRowNum(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-       */
-      public abstract void getRowNum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-      /**
-       * <code>rpc GetAvg(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-       */
-      public abstract void getAvg(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-      /**
-       * <code>rpc GetStd(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-       */
-      public abstract void getStd(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-      /**
-       * <code>rpc GetMedian(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-       */
-      public abstract void getMedian(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new AggregateService() {
-        @java.lang.Override
-        public  void getMax(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-          impl.getMax(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void getMin(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-          impl.getMin(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void getSum(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-          impl.getSum(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void getRowNum(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-          impl.getRowNum(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void getAvg(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-          impl.getAvg(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void getStd(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-          impl.getStd(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void getMedian(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-          impl.getMedian(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.getMax(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request);
-            case 1:
-              return impl.getMin(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request);
-            case 2:
-              return impl.getSum(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request);
-            case 3:
-              return impl.getRowNum(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request);
-            case 4:
-              return impl.getAvg(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request);
-            case 5:
-              return impl.getStd(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request);
-            case 6:
-              return impl.getMedian(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-            case 2:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-            case 3:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-            case 4:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-            case 5:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-            case 6:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-            case 2:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-            case 3:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-            case 4:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-            case 5:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-            case 6:
-              return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc GetMax(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-     */
-    public abstract void getMax(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-    /**
-     * <code>rpc GetMin(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-     */
-    public abstract void getMin(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-    /**
-     * <code>rpc GetSum(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-     */
-    public abstract void getSum(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-    /**
-     * <code>rpc GetRowNum(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-     */
-    public abstract void getRowNum(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-    /**
-     * <code>rpc GetAvg(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-     */
-    public abstract void getAvg(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-    /**
-     * <code>rpc GetStd(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-     */
-    public abstract void getStd(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-    /**
-     * <code>rpc GetMedian(.hbase.pb.AggregateRequest) returns (.hbase.pb.AggregateResponse);</code>
-     */
-    public abstract void getMedian(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.getMax(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback(
-              done));
-          return;
-        case 1:
-          this.getMin(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback(
-              done));
-          return;
-        case 2:
-          this.getSum(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback(
-              done));
-          return;
-        case 3:
-          this.getRowNum(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback(
-              done));
-          return;
-        case 4:
-          this.getAvg(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback(
-              done));
-          return;
-        case 5:
-          this.getStd(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback(
-              done));
-          return;
-        case 6:
-          this.getMedian(controller, (org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-        case 2:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-        case 3:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-        case 4:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-        case 5:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-        case 6:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-        case 2:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-        case 3:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-        case 4:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-        case 5:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-        case 6:
-          return org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateService implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void getMax(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.class,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance()));
-      }
-
-      public  void getMin(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.class,
-            org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance()));
-      }
-
-      public  void getSum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(2),
-          controller,
-          request,
-          org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.

<TRUNCATED>

[18/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MutabilityOracle.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MutabilityOracle.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MutabilityOracle.java
deleted file mode 100644
index b1eca4f..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MutabilityOracle.java
+++ /dev/null
@@ -1,48 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Verifies that an object is mutable, throwing if not.
- */
-interface MutabilityOracle {
-  static final MutabilityOracle IMMUTABLE = new MutabilityOracle() {
-    @Override
-    public void ensureMutable() {
-      throw new UnsupportedOperationException();
-    }
-  };
-
-  /**
-   * Throws an {@link UnsupportedOperationException} if not mutable.
-   */
-  void ensureMutable();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/NioByteString.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/NioByteString.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/NioByteString.java
deleted file mode 100644
index 1610f80..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/NioByteString.java
+++ /dev/null
@@ -1,291 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.checkNotNull;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.InvalidObjectException;
-import java.io.ObjectInputStream;
-import java.io.OutputStream;
-import java.nio.ByteBuffer;
-import java.nio.ByteOrder;
-import java.nio.InvalidMarkException;
-import java.nio.charset.Charset;
-import java.util.Collections;
-import java.util.List;
-
-/**
- * A {@link ByteString} that wraps around a {@link ByteBuffer}.
- */
-final class NioByteString extends ByteString.LeafByteString {
-  private final ByteBuffer buffer;
-
-  NioByteString(ByteBuffer buffer) {
-    checkNotNull(buffer, "buffer");
-
-    // Use native byte order for fast fixed32/64 operations.
-    this.buffer = buffer.slice().order(ByteOrder.nativeOrder());
-  }
-
-  // =================================================================
-  // Serializable
-
-  /**
-   * Magic method that lets us override serialization behavior.
-   */
-  private Object writeReplace() {
-    return ByteString.copyFrom(buffer.slice());
-  }
-
-  /**
-   * Magic method that lets us override deserialization behavior.
-   */
-  private void readObject(@SuppressWarnings("unused") ObjectInputStream in) throws IOException {
-    throw new InvalidObjectException("NioByteString instances are not to be serialized directly");
-  }
-
-  // =================================================================
-
-  @Override
-  public byte byteAt(int index) {
-    try {
-      return buffer.get(index);
-    } catch (ArrayIndexOutOfBoundsException e) {
-      throw e;
-    } catch (IndexOutOfBoundsException e) {
-      throw new ArrayIndexOutOfBoundsException(e.getMessage());
-    }
-  }
-
-  @Override
-  public int size() {
-    return buffer.remaining();
-  }
-
-  @Override
-  public ByteString substring(int beginIndex, int endIndex) {
-    try {
-      ByteBuffer slice = slice(beginIndex, endIndex);
-      return new NioByteString(slice);
-    } catch (ArrayIndexOutOfBoundsException e) {
-      throw e;
-    } catch (IndexOutOfBoundsException e) {
-      throw new ArrayIndexOutOfBoundsException(e.getMessage());
-    }
-  }
-
-  @Override
-  protected void copyToInternal(
-      byte[] target, int sourceOffset, int targetOffset, int numberToCopy) {
-    ByteBuffer slice = buffer.slice();
-    slice.position(sourceOffset);
-    slice.get(target, targetOffset, numberToCopy);
-  }
-
-  @Override
-  public void copyTo(ByteBuffer target) {
-    target.put(buffer.slice());
-  }
-
-  @Override
-  public void writeTo(OutputStream out) throws IOException {
-    out.write(toByteArray());
-  }
-
-  @Override
-  boolean equalsRange(ByteString other, int offset, int length) {
-    return substring(0, length).equals(other.substring(offset, offset + length));
-  }
-
-  @Override
-  void writeToInternal(OutputStream out, int sourceOffset, int numberToWrite) throws IOException {
-    if (buffer.hasArray()) {
-      // Optimized write for array-backed buffers.
-      // Note that we're taking the risk that a malicious OutputStream could modify the array.
-      int bufferOffset = buffer.arrayOffset() + buffer.position() + sourceOffset;
-      out.write(buffer.array(), bufferOffset, numberToWrite);
-      return;
-    }
-
-    ByteBufferWriter.write(slice(sourceOffset, sourceOffset + numberToWrite), out);
-  }
-
-  @Override
-  void writeTo(ByteOutput output) throws IOException {
-    output.writeLazy(buffer.slice());
-  }
-
-  @Override
-  public ByteBuffer asReadOnlyByteBuffer() {
-    return buffer.asReadOnlyBuffer();
-  }
-
-  @Override
-  public List<ByteBuffer> asReadOnlyByteBufferList() {
-    return Collections.singletonList(asReadOnlyByteBuffer());
-  }
-
-  @Override
-  protected String toStringInternal(Charset charset) {
-    final byte[] bytes;
-    final int offset;
-    final int length;
-    if (buffer.hasArray()) {
-      bytes = buffer.array();
-      offset = buffer.arrayOffset() + buffer.position();
-      length = buffer.remaining();
-    } else {
-      // TODO(nathanmittler): Can we optimize this?
-      bytes = toByteArray();
-      offset = 0;
-      length = bytes.length;
-    }
-    return new String(bytes, offset, length, charset);
-  }
-
-  @Override
-  public boolean isValidUtf8() {
-    return Utf8.isValidUtf8(buffer);
-  }
-
-  @Override
-  protected int partialIsValidUtf8(int state, int offset, int length) {
-    return Utf8.partialIsValidUtf8(state, buffer, offset, offset + length);
-  }
-
-  @Override
-  public boolean equals(Object other) {
-    if (other == this) {
-      return true;
-    }
-    if (!(other instanceof ByteString)) {
-      return false;
-    }
-    ByteString otherString = ((ByteString) other);
-    if (size() != otherString.size()) {
-      return false;
-    }
-    if (size() == 0) {
-      return true;
-    }
-    if (other instanceof NioByteString) {
-      return buffer.equals(((NioByteString) other).buffer);
-    }
-    if (other instanceof RopeByteString) {
-      return other.equals(this);
-    }
-    return buffer.equals(otherString.asReadOnlyByteBuffer());
-  }
-
-  @Override
-  protected int partialHash(int h, int offset, int length) {
-    for (int i = offset; i < offset + length; i++) {
-      h = h * 31 + buffer.get(i);
-    }
-    return h;
-  }
-
-  @Override
-  public InputStream newInput() {
-    return new InputStream() {
-      private final ByteBuffer buf = buffer.slice();
-
-      @Override
-      public void mark(int readlimit) {
-        buf.mark();
-      }
-
-      @Override
-      public boolean markSupported() {
-        return true;
-      }
-
-      @Override
-      public void reset() throws IOException {
-        try {
-          buf.reset();
-        } catch (InvalidMarkException e) {
-          throw new IOException(e);
-        }
-      }
-
-      @Override
-      public int available() throws IOException {
-        return buf.remaining();
-      }
-
-      @Override
-      public int read() throws IOException {
-        if (!buf.hasRemaining()) {
-          return -1;
-        }
-        return buf.get() & 0xFF;
-      }
-
-      @Override
-      public int read(byte[] bytes, int off, int len) throws IOException {
-        if (!buf.hasRemaining()) {
-          return -1;
-        }
-
-        len = Math.min(len, buf.remaining());
-        buf.get(bytes, off, len);
-        return len;
-      }
-    };
-  }
-
-  @Override
-  public CodedInputStream newCodedInput() {
-    return CodedInputStream.newInstance(buffer, true);
-  }
-
-  /**
-   * Creates a slice of a range of this buffer.
-   *
-   * @param beginIndex the beginning index of the slice (inclusive).
-   * @param endIndex the end index of the slice (exclusive).
-   * @return the requested slice.
-   */
-  private ByteBuffer slice(int beginIndex, int endIndex) {
-    if (beginIndex < buffer.position() || endIndex > buffer.limit() || beginIndex > endIndex) {
-      throw new IllegalArgumentException(
-          String.format("Invalid indices [%d, %d]", beginIndex, endIndex));
-    }
-
-    ByteBuffer slice = buffer.slice();
-    slice.position(beginIndex - buffer.position());
-    slice.limit(endIndex - buffer.position());
-    return slice;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/NullValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/NullValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/NullValue.java
deleted file mode 100644
index 6cedb68..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/NullValue.java
+++ /dev/null
@@ -1,108 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/struct.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * `NullValue` is a singleton enumeration to represent the null value for the
- * `Value` type union.
- *  The JSON representation for `NullValue` is JSON `null`.
- * </pre>
- *
- * Protobuf enum {@code google.protobuf.NullValue}
- */
-public enum NullValue
-    implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
-  /**
-   * <pre>
-   * Null value.
-   * </pre>
-   *
-   * <code>NULL_VALUE = 0;</code>
-   */
-  NULL_VALUE(0),
-  UNRECOGNIZED(-1),
-  ;
-
-  /**
-   * <pre>
-   * Null value.
-   * </pre>
-   *
-   * <code>NULL_VALUE = 0;</code>
-   */
-  public static final int NULL_VALUE_VALUE = 0;
-
-
-  public final int getNumber() {
-    if (this == UNRECOGNIZED) {
-      throw new java.lang.IllegalArgumentException(
-          "Can't get the number of an unknown enum value.");
-    }
-    return value;
-  }
-
-  /**
-   * @deprecated Use {@link #forNumber(int)} instead.
-   */
-  @java.lang.Deprecated
-  public static NullValue valueOf(int value) {
-    return forNumber(value);
-  }
-
-  public static NullValue forNumber(int value) {
-    switch (value) {
-      case 0: return NULL_VALUE;
-      default: return null;
-    }
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<NullValue>
-      internalGetValueMap() {
-    return internalValueMap;
-  }
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
-      NullValue> internalValueMap =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<NullValue>() {
-          public NullValue findValueByNumber(int number) {
-            return NullValue.forNumber(number);
-          }
-        };
-
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
-      getValueDescriptor() {
-    return getDescriptor().getValues().get(ordinal());
-  }
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-      getDescriptorForType() {
-    return getDescriptor();
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.getDescriptor().getEnumTypes().get(0);
-  }
-
-  private static final NullValue[] VALUES = values();
-
-  public static NullValue valueOf(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-    if (desc.getType() != getDescriptor()) {
-      throw new java.lang.IllegalArgumentException(
-        "EnumValueDescriptor is not for this type.");
-    }
-    if (desc.getIndex() == -1) {
-      return UNRECOGNIZED;
-    }
-    return VALUES[desc.getIndex()];
-  }
-
-  private final int value;
-
-  private NullValue(int value) {
-    this.value = value;
-  }
-
-  // @@protoc_insertion_point(enum_scope:google.protobuf.NullValue)
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Option.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Option.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Option.java
deleted file mode 100644
index cfd28c9..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Option.java
+++ /dev/null
@@ -1,818 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * A protocol buffer option, which can be attached to a message, field,
- * enumeration, etc.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Option}
- */
-public  final class Option extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Option)
-    OptionOrBuilder {
-  // Use Option.newBuilder() to construct.
-  private Option(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Option() {
-    name_ = "";
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Option(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            name_ = s;
-            break;
-          }
-          case 18: {
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.Builder subBuilder = null;
-            if (value_ != null) {
-              subBuilder = value_.toBuilder();
-            }
-            value_ = input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.parser(), extensionRegistry);
-            if (subBuilder != null) {
-              subBuilder.mergeFrom(value_);
-              value_ = subBuilder.buildPartial();
-            }
-
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Option_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Option_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder.class);
-  }
-
-  public static final int NAME_FIELD_NUMBER = 1;
-  private volatile java.lang.Object name_;
-  /**
-   * <pre>
-   * The option's name. For protobuf built-in options (options defined in
-   * descriptor.proto), this is the short name. For example, `"map_entry"`.
-   * For custom options, it should be the fully-qualified name. For example,
-   * `"google.api.http"`.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public java.lang.String getName() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      name_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The option's name. For protobuf built-in options (options defined in
-   * descriptor.proto), this is the short name. For example, `"map_entry"`.
-   * For custom options, it should be the fully-qualified name. For example,
-   * `"google.api.http"`.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      name_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 2;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.Any value_;
-  /**
-   * <pre>
-   * The option's value packed in an Any message. If the value is a primitive,
-   * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-   * should be used. If the value is an enum, it should be stored as an int32
-   * value using the google.protobuf.Int32Value type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Any value = 2;</code>
-   */
-  public boolean hasValue() {
-    return value_ != null;
-  }
-  /**
-   * <pre>
-   * The option's value packed in an Any message. If the value is a primitive,
-   * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-   * should be used. If the value is an enum, it should be stored as an int32
-   * value using the google.protobuf.Int32Value type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Any value = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Any getValue() {
-    return value_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.getDefaultInstance() : value_;
-  }
-  /**
-   * <pre>
-   * The option's value packed in an Any message. If the value is a primitive,
-   * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-   * should be used. If the value is an enum, it should be stored as an int32
-   * value using the google.protobuf.Int32Value type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Any value = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyOrBuilder getValueOrBuilder() {
-    return getValue();
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
-    }
-    if (value_ != null) {
-      output.writeMessage(2, getValue());
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
-    }
-    if (value_ != null) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(2, getValue());
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Option)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Option other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Option) obj;
-
-    boolean result = true;
-    result = result && getName()
-        .equals(other.getName());
-    result = result && (hasValue() == other.hasValue());
-    if (hasValue()) {
-      result = result && getValue()
-          .equals(other.getValue());
-    }
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getName().hashCode();
-    if (hasValue()) {
-      hash = (37 * hash) + VALUE_FIELD_NUMBER;
-      hash = (53 * hash) + getValue().hashCode();
-    }
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * A protocol buffer option, which can be attached to a message, field,
-   * enumeration, etc.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Option}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Option)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Option_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Option_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      name_ = "";
-
-      if (valueBuilder_ == null) {
-        value_ = null;
-      } else {
-        value_ = null;
-        valueBuilder_ = null;
-      }
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Option_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Option result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Option result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Option(this);
-      result.name_ = name_;
-      if (valueBuilder_ == null) {
-        result.value_ = value_;
-      } else {
-        result.value_ = valueBuilder_.build();
-      }
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Option) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Option)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance()) return this;
-      if (!other.getName().isEmpty()) {
-        name_ = other.name_;
-        onChanged();
-      }
-      if (other.hasValue()) {
-        mergeValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Option parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Option) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private java.lang.Object name_ = "";
-    /**
-     * <pre>
-     * The option's name. For protobuf built-in options (options defined in
-     * descriptor.proto), this is the short name. For example, `"map_entry"`.
-     * For custom options, it should be the fully-qualified name. For example,
-     * `"google.api.http"`.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        name_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The option's name. For protobuf built-in options (options defined in
-     * descriptor.proto), this is the short name. For example, `"map_entry"`.
-     * For custom options, it should be the fully-qualified name. For example,
-     * `"google.api.http"`.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The option's name. For protobuf built-in options (options defined in
-     * descriptor.proto), this is the short name. For example, `"map_entry"`.
-     * For custom options, it should be the fully-qualified name. For example,
-     * `"google.api.http"`.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      name_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The option's name. For protobuf built-in options (options defined in
-     * descriptor.proto), this is the short name. For example, `"map_entry"`.
-     * For custom options, it should be the fully-qualified name. For example,
-     * `"google.api.http"`.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder clearName() {
-      
-      name_ = getDefaultInstance().getName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The option's name. For protobuf built-in options (options defined in
-     * descriptor.proto), this is the short name. For example, `"map_entry"`.
-     * For custom options, it should be the fully-qualified name. For example,
-     * `"google.api.http"`.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      name_ = value;
-      onChanged();
-      return this;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.Any value_ = null;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Any, org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyOrBuilder> valueBuilder_;
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    public boolean hasValue() {
-      return valueBuilder_ != null || value_ != null;
-    }
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Any getValue() {
-      if (valueBuilder_ == null) {
-        return value_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.getDefaultInstance() : value_;
-      } else {
-        return valueBuilder_.getMessage();
-      }
-    }
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    public Builder setValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.Any value) {
-      if (valueBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        value_ = value;
-        onChanged();
-      } else {
-        valueBuilder_.setMessage(value);
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    public Builder setValue(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.Builder builderForValue) {
-      if (valueBuilder_ == null) {
-        value_ = builderForValue.build();
-        onChanged();
-      } else {
-        valueBuilder_.setMessage(builderForValue.build());
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    public Builder mergeValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.Any value) {
-      if (valueBuilder_ == null) {
-        if (value_ != null) {
-          value_ =
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.newBuilder(value_).mergeFrom(value).buildPartial();
-        } else {
-          value_ = value;
-        }
-        onChanged();
-      } else {
-        valueBuilder_.mergeFrom(value);
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    public Builder clearValue() {
-      if (valueBuilder_ == null) {
-        value_ = null;
-        onChanged();
-      } else {
-        value_ = null;
-        valueBuilder_ = null;
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.Builder getValueBuilder() {
-      
-      onChanged();
-      return getValueFieldBuilder().getBuilder();
-    }
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyOrBuilder getValueOrBuilder() {
-      if (valueBuilder_ != null) {
-        return valueBuilder_.getMessageOrBuilder();
-      } else {
-        return value_ == null ?
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.getDefaultInstance() : value_;
-      }
-    }
-    /**
-     * <pre>
-     * The option's value packed in an Any message. If the value is a primitive,
-     * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-     * should be used. If the value is an enum, it should be stored as an int32
-     * value using the google.protobuf.Int32Value type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Any value = 2;</code>
-     */
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Any, org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyOrBuilder> 
-        getValueFieldBuilder() {
-      if (valueBuilder_ == null) {
-        valueBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Any, org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyOrBuilder>(
-                getValue(),
-                getParentForChildren(),
-                isClean());
-        value_ = null;
-      }
-      return valueBuilder_;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Option)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Option)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Option DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Option();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Option>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Option>() {
-    public Option parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Option(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Option> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Option> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/OptionOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/OptionOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/OptionOrBuilder.java
deleted file mode 100644
index 3fea10a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/OptionOrBuilder.java
+++ /dev/null
@@ -1,67 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface OptionOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Option)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The option's name. For protobuf built-in options (options defined in
-   * descriptor.proto), this is the short name. For example, `"map_entry"`.
-   * For custom options, it should be the fully-qualified name. For example,
-   * `"google.api.http"`.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  java.lang.String getName();
-  /**
-   * <pre>
-   * The option's name. For protobuf built-in options (options defined in
-   * descriptor.proto), this is the short name. For example, `"map_entry"`.
-   * For custom options, it should be the fully-qualified name. For example,
-   * `"google.api.http"`.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes();
-
-  /**
-   * <pre>
-   * The option's value packed in an Any message. If the value is a primitive,
-   * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-   * should be used. If the value is an enum, it should be stored as an int32
-   * value using the google.protobuf.Int32Value type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Any value = 2;</code>
-   */
-  boolean hasValue();
-  /**
-   * <pre>
-   * The option's value packed in an Any message. If the value is a primitive,
-   * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-   * should be used. If the value is an enum, it should be stored as an int32
-   * value using the google.protobuf.Int32Value type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Any value = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Any getValue();
-  /**
-   * <pre>
-   * The option's value packed in an Any message. If the value is a primitive,
-   * the corresponding wrapper type defined in google/protobuf/wrappers.proto
-   * should be used. If the value is an enum, it should be stored as an int32
-   * value using the google.protobuf.Int32Value type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Any value = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyOrBuilder getValueOrBuilder();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Parser.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Parser.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Parser.java
deleted file mode 100644
index f29b60b..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Parser.java
+++ /dev/null
@@ -1,272 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.InputStream;
-
-/**
- * Abstract interface for parsing Protocol Messages.
- *
- * The implementation should be stateless and thread-safe.
- *
- * <p>All methods may throw {@link InvalidProtocolBufferException}. In the event of invalid data,
- * like an encoding error, the cause of the thrown exception will be {@code null}. However, if an
- * I/O problem occurs, an exception is thrown with an {@link java.io.IOException} cause.
- *
- * @author liujisi@google.com (Pherl Liu)
- */
-public interface Parser<MessageType> {
-
-  // NB(jh): Other parts of the protobuf API that parse messages distinguish between an I/O problem
-  // (like failure reading bytes from a socket) and invalid data (encoding error) via the type of
-  // thrown exception. But it would be source-incompatible to make the methods in this interface do
-  // so since they were originally spec'ed to only throw InvalidProtocolBufferException. So callers
-  // must inspect the cause of the exception to distinguish these two cases.
-
-  /**
-   * Parses a message of {@code MessageType} from the input.
-   *
-   * <p>Note:  The caller should call
-   * {@link CodedInputStream#checkLastTagWas(int)} after calling this to
-   * verify that the last tag seen was the appropriate end-group tag,
-   * or zero for EOF.
-   */
-  public MessageType parseFrom(CodedInputStream input)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(CodedInputStream)}, but also parses extensions.
-   * The extensions that you want to be able to parse must be registered in
-   * {@code extensionRegistry}. Extensions not in the registry will be treated
-   * as unknown fields.
-   */
-  public MessageType parseFrom(CodedInputStream input,
-                               ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(CodedInputStream)}, but does not throw an
-   * exception if the message is missing required fields. Instead, a partial
-   * message is returned.
-   */
-  public MessageType parsePartialFrom(CodedInputStream input)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(CodedInputStream input, ExtensionRegistryLite)},
-   * but does not throw an exception if the message is missing required fields.
-   * Instead, a partial message is returned.
-   */
-  public MessageType parsePartialFrom(CodedInputStream input,
-                                      ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  // ---------------------------------------------------------------
-  // Convenience methods.
-
-  /**
-   * Parses {@code data} as a message of {@code MessageType}.
-   * This is just a small wrapper around {@link #parseFrom(CodedInputStream)}.
-   */
-  public MessageType parseFrom(ByteString data)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Parses {@code data} as a message of {@code MessageType}.
-   * This is just a small wrapper around
-   * {@link #parseFrom(CodedInputStream, ExtensionRegistryLite)}.
-   */
-  public MessageType parseFrom(ByteString data,
-                               ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(ByteString)}, but does not throw an
-   * exception if the message is missing required fields. Instead, a partial
-   * message is returned.
-   */
-  public MessageType parsePartialFrom(ByteString data)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(ByteString, ExtensionRegistryLite)},
-   * but does not throw an exception if the message is missing required fields.
-   * Instead, a partial message is returned.
-   */
-  public MessageType parsePartialFrom(ByteString data,
-                                      ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Parses {@code data} as a message of {@code MessageType}.
-   * This is just a small wrapper around {@link #parseFrom(CodedInputStream)}.
-   */
-  public MessageType parseFrom(byte[] data, int off, int len)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Parses {@code data} as a message of {@code MessageType}.
-   * This is just a small wrapper around
-   * {@link #parseFrom(CodedInputStream, ExtensionRegistryLite)}.
-   */
-  public MessageType parseFrom(byte[] data, int off, int len,
-                               ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Parses {@code data} as a message of {@code MessageType}.
-   * This is just a small wrapper around {@link #parseFrom(CodedInputStream)}.
-   */
-  public MessageType parseFrom(byte[] data)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Parses {@code data} as a message of {@code MessageType}.
-   * This is just a small wrapper around
-   * {@link #parseFrom(CodedInputStream, ExtensionRegistryLite)}.
-   */
-  public MessageType parseFrom(byte[] data,
-                               ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(byte[], int, int)}, but does not throw an
-   * exception if the message is missing required fields. Instead, a partial
-   * message is returned.
-   */
-  public MessageType parsePartialFrom(byte[] data, int off, int len)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(ByteString, ExtensionRegistryLite)},
-   * but does not throw an exception if the message is missing required fields.
-   * Instead, a partial message is returned.
-   */
-  public MessageType parsePartialFrom(byte[] data, int off, int len,
-                                      ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(byte[])}, but does not throw an
-   * exception if the message is missing required fields. Instead, a partial
-   * message is returned.
-   */
-  public MessageType parsePartialFrom(byte[] data)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(byte[], ExtensionRegistryLite)},
-   * but does not throw an exception if the message is missing required fields.
-   * Instead, a partial message is returned.
-   */
-  public MessageType parsePartialFrom(byte[] data,
-                                      ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Parse a message of {@code MessageType} from {@code input}.
-   * This is just a small wrapper around {@link #parseFrom(CodedInputStream)}.
-   * Note that this method always reads the <i>entire</i> input (unless it
-   * throws an exception).  If you want it to stop earlier, you will need to
-   * wrap your input in some wrapper stream that limits reading.  Or, use
-   * {@link MessageLite#writeDelimitedTo(java.io.OutputStream)} to write your
-   * message and {@link #parseDelimitedFrom(InputStream)} to read it.
-   * <p>
-   * Despite usually reading the entire input, this does not close the stream.
-   */
-  public MessageType parseFrom(InputStream input)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Parses a message of {@code MessageType} from {@code input}.
-   * This is just a small wrapper around
-   * {@link #parseFrom(CodedInputStream, ExtensionRegistryLite)}.
-   */
-  public MessageType parseFrom(InputStream input,
-                               ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(InputStream)}, but does not throw an
-   * exception if the message is missing required fields. Instead, a partial
-   * message is returned.
-   */
-  public MessageType parsePartialFrom(InputStream input)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(InputStream, ExtensionRegistryLite)},
-   * but does not throw an exception if the message is missing required fields.
-   * Instead, a partial message is returned.
-   */
-  public MessageType parsePartialFrom(InputStream input,
-                                      ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseFrom(InputStream)}, but does not read util EOF.
-   * Instead, the size of message (encoded as a varint) is read first,
-   * then the message data. Use
-   * {@link MessageLite#writeDelimitedTo(java.io.OutputStream)} to write
-   * messages in this format.
-   *
-   * @return Parsed message if successful, or null if the stream is at EOF when
-   *         the method starts. Any other error (including reaching EOF during
-   *         parsing) will cause an exception to be thrown.
-   */
-  public MessageType parseDelimitedFrom(InputStream input)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseDelimitedFrom(InputStream)} but supporting extensions.
-   */
-  public MessageType parseDelimitedFrom(InputStream input,
-                                        ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseDelimitedFrom(InputStream)}, but does not throw an
-   * exception if the message is missing required fields. Instead, a partial
-   * message is returned.
-   */
-  public MessageType parsePartialDelimitedFrom(InputStream input)
-      throws InvalidProtocolBufferException;
-
-  /**
-   * Like {@link #parseDelimitedFrom(InputStream, ExtensionRegistryLite)},
-   * but does not throw an exception if the message is missing required fields.
-   * Instead, a partial message is returned.
-   */
-  public MessageType parsePartialDelimitedFrom(
-      InputStream input,
-      ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException;
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtobufArrayList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtobufArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtobufArrayList.java
deleted file mode 100644
index 0f6ecd1..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtobufArrayList.java
+++ /dev/null
@@ -1,105 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.ProtobufList;
-
-import java.util.ArrayList;
-import java.util.List;
-
-/**
- * Implements {@link ProtobufList} for non-primitive and {@link String} types.
- */
-final class ProtobufArrayList<E> extends AbstractProtobufList<E> {
-
-  private static final ProtobufArrayList<Object> EMPTY_LIST = new ProtobufArrayList<Object>();
-  static {
-    EMPTY_LIST.makeImmutable();
-  }
-  
-  @SuppressWarnings("unchecked") // Guaranteed safe by runtime.
-  public static <E> ProtobufArrayList<E> emptyList() {
-    return (ProtobufArrayList<E>) EMPTY_LIST;
-  }
-  
-  private final List<E> list;
-
-  ProtobufArrayList() {
-    this(new ArrayList<E>(DEFAULT_CAPACITY));
-  }
-  
-  private ProtobufArrayList(List<E> list) {
-    this.list = list;
-  }
-
-  @Override
-  public ProtobufArrayList<E> mutableCopyWithCapacity(int capacity) {
-    if (capacity < size()) {
-      throw new IllegalArgumentException();
-    }
-    List<E> newList = new ArrayList<E>(capacity);
-    newList.addAll(list);
-    return new ProtobufArrayList<E>(newList);
-  }
-  
-  @Override
-  public void add(int index, E element) {
-    ensureIsMutable();
-    list.add(index, element);
-    modCount++;
-  }
-
-  @Override
-  public E get(int index) {
-    return list.get(index);
-  }
-  
-  @Override
-  public E remove(int index) {
-    ensureIsMutable();
-    E toReturn = list.remove(index);
-    modCount++;
-    return toReturn;
-  }
-  
-  @Override
-  public E set(int index, E element) {
-    ensureIsMutable();
-    E toReturn = list.set(index, element);
-    modCount++;
-    return toReturn;
-  }
-
-  @Override
-  public int size() {
-    return list.size();
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtocolMessageEnum.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtocolMessageEnum.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtocolMessageEnum.java
deleted file mode 100644
index 7b40fbe..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtocolMessageEnum.java
+++ /dev/null
@@ -1,59 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor;
-
-/**
- * Interface of useful methods added to all enums generated by the protocol
- * compiler.
- */
-public interface ProtocolMessageEnum extends Internal.EnumLite {
-
-  /**
-   * Return the value's numeric value as defined in the .proto file.
-   */
-  @Override
-  int getNumber();
-
-  /**
-   * Return the value's descriptor, which contains information such as
-   * value name, number, and type.
-   */
-  EnumValueDescriptor getValueDescriptor();
-
-  /**
-   * Return the enum type's descriptor, which contains information
-   * about each defined value, etc.
-   */
-  EnumDescriptor getDescriptorForType();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtocolStringList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtocolStringList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtocolStringList.java
deleted file mode 100644
index 6c08888..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ProtocolStringList.java
+++ /dev/null
@@ -1,48 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.List;
-
-/**
- * An interface extending {@code List<String>} used for repeated string fields
- * to provide optional access to the data as a list of ByteStrings. The
- * underlying implementation stores values as either ByteStrings or Strings
- * (see {@link LazyStringArrayList}) depending on how the value was initialized
- * or last read, and it is often more efficient to deal with lists of
- * ByteStrings when handling protos that have been deserialized from bytes.
- */
-public interface ProtocolStringList extends List<String> {
-
-  /** Returns a view of the data as a list of ByteStrings. */
-  List<ByteString> asByteStringList();
-
-}


[12/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Type.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Type.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Type.java
deleted file mode 100644
index f46f351..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Type.java
+++ /dev/null
@@ -1,1967 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * A protocol buffer message type.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Type}
- */
-public  final class Type extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Type)
-    TypeOrBuilder {
-  // Use Type.newBuilder() to construct.
-  private Type(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Type() {
-    name_ = "";
-    fields_ = java.util.Collections.emptyList();
-    oneofs_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-    options_ = java.util.Collections.emptyList();
-    syntax_ = 0;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Type(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            name_ = s;
-            break;
-          }
-          case 18: {
-            if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-              fields_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Field>();
-              mutable_bitField0_ |= 0x00000002;
-            }
-            fields_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.parser(), extensionRegistry));
-            break;
-          }
-          case 26: {
-            java.lang.String s = input.readStringRequireUtf8();
-            if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-              oneofs_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList();
-              mutable_bitField0_ |= 0x00000004;
-            }
-            oneofs_.add(s);
-            break;
-          }
-          case 34: {
-            if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-              options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>();
-              mutable_bitField0_ |= 0x00000008;
-            }
-            options_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.parser(), extensionRegistry));
-            break;
-          }
-          case 42: {
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder subBuilder = null;
-            if (sourceContext_ != null) {
-              subBuilder = sourceContext_.toBuilder();
-            }
-            sourceContext_ = input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.parser(), extensionRegistry);
-            if (subBuilder != null) {
-              subBuilder.mergeFrom(sourceContext_);
-              sourceContext_ = subBuilder.buildPartial();
-            }
-
-            break;
-          }
-          case 48: {
-            int rawValue = input.readEnum();
-
-            syntax_ = rawValue;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-        fields_ = java.util.Collections.unmodifiableList(fields_);
-      }
-      if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-        oneofs_ = oneofs_.getUnmodifiableView();
-      }
-      if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-        options_ = java.util.Collections.unmodifiableList(options_);
-      }
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Type_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Type_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Type.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Type.Builder.class);
-  }
-
-  private int bitField0_;
-  public static final int NAME_FIELD_NUMBER = 1;
-  private volatile java.lang.Object name_;
-  /**
-   * <pre>
-   * The fully qualified message name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public java.lang.String getName() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      name_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The fully qualified message name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      name_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int FIELDS_FIELD_NUMBER = 2;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Field> fields_;
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Field> getFieldsList() {
-    return fields_;
-  }
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder> 
-      getFieldsOrBuilderList() {
-    return fields_;
-  }
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  public int getFieldsCount() {
-    return fields_.size();
-  }
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field getFields(int index) {
-    return fields_.get(index);
-  }
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder getFieldsOrBuilder(
-      int index) {
-    return fields_.get(index);
-  }
-
-  public static final int ONEOFS_FIELD_NUMBER = 3;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList oneofs_;
-  /**
-   * <pre>
-   * The list of types appearing in `oneof` definitions in this type.
-   * </pre>
-   *
-   * <code>repeated string oneofs = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
-      getOneofsList() {
-    return oneofs_;
-  }
-  /**
-   * <pre>
-   * The list of types appearing in `oneof` definitions in this type.
-   * </pre>
-   *
-   * <code>repeated string oneofs = 3;</code>
-   */
-  public int getOneofsCount() {
-    return oneofs_.size();
-  }
-  /**
-   * <pre>
-   * The list of types appearing in `oneof` definitions in this type.
-   * </pre>
-   *
-   * <code>repeated string oneofs = 3;</code>
-   */
-  public java.lang.String getOneofs(int index) {
-    return oneofs_.get(index);
-  }
-  /**
-   * <pre>
-   * The list of types appearing in `oneof` definitions in this type.
-   * </pre>
-   *
-   * <code>repeated string oneofs = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getOneofsBytes(int index) {
-    return oneofs_.getByteString(index);
-  }
-
-  public static final int OPTIONS_FIELD_NUMBER = 4;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_;
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  public int getOptionsCount() {
-    return options_.size();
-  }
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-    return options_.get(index);
-  }
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index) {
-    return options_.get(index);
-  }
-
-  public static final int SOURCE_CONTEXT_FIELD_NUMBER = 5;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext sourceContext_;
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  public boolean hasSourceContext() {
-    return sourceContext_ != null;
-  }
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext() {
-    return sourceContext_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-  }
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder() {
-    return getSourceContext();
-  }
-
-  public static final int SYNTAX_FIELD_NUMBER = 6;
-  private int syntax_;
-  /**
-   * <pre>
-   * The source syntax.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 6;</code>
-   */
-  public int getSyntaxValue() {
-    return syntax_;
-  }
-  /**
-   * <pre>
-   * The source syntax.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 6;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax() {
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.valueOf(syntax_);
-    return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.UNRECOGNIZED : result;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
-    }
-    for (int i = 0; i < fields_.size(); i++) {
-      output.writeMessage(2, fields_.get(i));
-    }
-    for (int i = 0; i < oneofs_.size(); i++) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, oneofs_.getRaw(i));
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      output.writeMessage(4, options_.get(i));
-    }
-    if (sourceContext_ != null) {
-      output.writeMessage(5, getSourceContext());
-    }
-    if (syntax_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.SYNTAX_PROTO2.getNumber()) {
-      output.writeEnum(6, syntax_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
-    }
-    for (int i = 0; i < fields_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(2, fields_.get(i));
-    }
-    {
-      int dataSize = 0;
-      for (int i = 0; i < oneofs_.size(); i++) {
-        dataSize += computeStringSizeNoTag(oneofs_.getRaw(i));
-      }
-      size += dataSize;
-      size += 1 * getOneofsList().size();
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(4, options_.get(i));
-    }
-    if (sourceContext_ != null) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(5, getSourceContext());
-    }
-    if (syntax_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.SYNTAX_PROTO2.getNumber()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeEnumSize(6, syntax_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Type)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Type other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Type) obj;
-
-    boolean result = true;
-    result = result && getName()
-        .equals(other.getName());
-    result = result && getFieldsList()
-        .equals(other.getFieldsList());
-    result = result && getOneofsList()
-        .equals(other.getOneofsList());
-    result = result && getOptionsList()
-        .equals(other.getOptionsList());
-    result = result && (hasSourceContext() == other.hasSourceContext());
-    if (hasSourceContext()) {
-      result = result && getSourceContext()
-          .equals(other.getSourceContext());
-    }
-    result = result && syntax_ == other.syntax_;
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getName().hashCode();
-    if (getFieldsCount() > 0) {
-      hash = (37 * hash) + FIELDS_FIELD_NUMBER;
-      hash = (53 * hash) + getFieldsList().hashCode();
-    }
-    if (getOneofsCount() > 0) {
-      hash = (37 * hash) + ONEOFS_FIELD_NUMBER;
-      hash = (53 * hash) + getOneofsList().hashCode();
-    }
-    if (getOptionsCount() > 0) {
-      hash = (37 * hash) + OPTIONS_FIELD_NUMBER;
-      hash = (53 * hash) + getOptionsList().hashCode();
-    }
-    if (hasSourceContext()) {
-      hash = (37 * hash) + SOURCE_CONTEXT_FIELD_NUMBER;
-      hash = (53 * hash) + getSourceContext().hashCode();
-    }
-    hash = (37 * hash) + SYNTAX_FIELD_NUMBER;
-    hash = (53 * hash) + syntax_;
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Type prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * A protocol buffer message type.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Type}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Type)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Type_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Type_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Type.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Type.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Type.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-        getFieldsFieldBuilder();
-        getOptionsFieldBuilder();
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      name_ = "";
-
-      if (fieldsBuilder_ == null) {
-        fields_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000002);
-      } else {
-        fieldsBuilder_.clear();
-      }
-      oneofs_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-      bitField0_ = (bitField0_ & ~0x00000004);
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000008);
-      } else {
-        optionsBuilder_.clear();
-      }
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = null;
-      } else {
-        sourceContext_ = null;
-        sourceContextBuilder_ = null;
-      }
-      syntax_ = 0;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Type_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Type getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Type.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Type build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Type result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Type buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Type result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Type(this);
-      int from_bitField0_ = bitField0_;
-      int to_bitField0_ = 0;
-      result.name_ = name_;
-      if (fieldsBuilder_ == null) {
-        if (((bitField0_ & 0x00000002) == 0x00000002)) {
-          fields_ = java.util.Collections.unmodifiableList(fields_);
-          bitField0_ = (bitField0_ & ~0x00000002);
-        }
-        result.fields_ = fields_;
-      } else {
-        result.fields_ = fieldsBuilder_.build();
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        oneofs_ = oneofs_.getUnmodifiableView();
-        bitField0_ = (bitField0_ & ~0x00000004);
-      }
-      result.oneofs_ = oneofs_;
-      if (optionsBuilder_ == null) {
-        if (((bitField0_ & 0x00000008) == 0x00000008)) {
-          options_ = java.util.Collections.unmodifiableList(options_);
-          bitField0_ = (bitField0_ & ~0x00000008);
-        }
-        result.options_ = options_;
-      } else {
-        result.options_ = optionsBuilder_.build();
-      }
-      if (sourceContextBuilder_ == null) {
-        result.sourceContext_ = sourceContext_;
-      } else {
-        result.sourceContext_ = sourceContextBuilder_.build();
-      }
-      result.syntax_ = syntax_;
-      result.bitField0_ = to_bitField0_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Type) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Type)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Type other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Type.getDefaultInstance()) return this;
-      if (!other.getName().isEmpty()) {
-        name_ = other.name_;
-        onChanged();
-      }
-      if (fieldsBuilder_ == null) {
-        if (!other.fields_.isEmpty()) {
-          if (fields_.isEmpty()) {
-            fields_ = other.fields_;
-            bitField0_ = (bitField0_ & ~0x00000002);
-          } else {
-            ensureFieldsIsMutable();
-            fields_.addAll(other.fields_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.fields_.isEmpty()) {
-          if (fieldsBuilder_.isEmpty()) {
-            fieldsBuilder_.dispose();
-            fieldsBuilder_ = null;
-            fields_ = other.fields_;
-            bitField0_ = (bitField0_ & ~0x00000002);
-            fieldsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getFieldsFieldBuilder() : null;
-          } else {
-            fieldsBuilder_.addAllMessages(other.fields_);
-          }
-        }
-      }
-      if (!other.oneofs_.isEmpty()) {
-        if (oneofs_.isEmpty()) {
-          oneofs_ = other.oneofs_;
-          bitField0_ = (bitField0_ & ~0x00000004);
-        } else {
-          ensureOneofsIsMutable();
-          oneofs_.addAll(other.oneofs_);
-        }
-        onChanged();
-      }
-      if (optionsBuilder_ == null) {
-        if (!other.options_.isEmpty()) {
-          if (options_.isEmpty()) {
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000008);
-          } else {
-            ensureOptionsIsMutable();
-            options_.addAll(other.options_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.options_.isEmpty()) {
-          if (optionsBuilder_.isEmpty()) {
-            optionsBuilder_.dispose();
-            optionsBuilder_ = null;
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000008);
-            optionsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getOptionsFieldBuilder() : null;
-          } else {
-            optionsBuilder_.addAllMessages(other.options_);
-          }
-        }
-      }
-      if (other.hasSourceContext()) {
-        mergeSourceContext(other.getSourceContext());
-      }
-      if (other.syntax_ != 0) {
-        setSyntaxValue(other.getSyntaxValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Type parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Type) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private java.lang.Object name_ = "";
-    /**
-     * <pre>
-     * The fully qualified message name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        name_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The fully qualified message name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The fully qualified message name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      name_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The fully qualified message name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder clearName() {
-      
-      name_ = getDefaultInstance().getName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The fully qualified message name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      name_ = value;
-      onChanged();
-      return this;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Field> fields_ =
-      java.util.Collections.emptyList();
-    private void ensureFieldsIsMutable() {
-      if (!((bitField0_ & 0x00000002) == 0x00000002)) {
-        fields_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Field>(fields_);
-        bitField0_ |= 0x00000002;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Field, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder> fieldsBuilder_;
-
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Field> getFieldsList() {
-      if (fieldsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(fields_);
-      } else {
-        return fieldsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public int getFieldsCount() {
-      if (fieldsBuilder_ == null) {
-        return fields_.size();
-      } else {
-        return fieldsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field getFields(int index) {
-      if (fieldsBuilder_ == null) {
-        return fields_.get(index);
-      } else {
-        return fieldsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder setFields(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field value) {
-      if (fieldsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureFieldsIsMutable();
-        fields_.set(index, value);
-        onChanged();
-      } else {
-        fieldsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder setFields(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder builderForValue) {
-      if (fieldsBuilder_ == null) {
-        ensureFieldsIsMutable();
-        fields_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        fieldsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder addFields(org.apache.hadoop.hbase.shaded.com.google.protobuf.Field value) {
-      if (fieldsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureFieldsIsMutable();
-        fields_.add(value);
-        onChanged();
-      } else {
-        fieldsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder addFields(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field value) {
-      if (fieldsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureFieldsIsMutable();
-        fields_.add(index, value);
-        onChanged();
-      } else {
-        fieldsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder addFields(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder builderForValue) {
-      if (fieldsBuilder_ == null) {
-        ensureFieldsIsMutable();
-        fields_.add(builderForValue.build());
-        onChanged();
-      } else {
-        fieldsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder addFields(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder builderForValue) {
-      if (fieldsBuilder_ == null) {
-        ensureFieldsIsMutable();
-        fields_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        fieldsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder addAllFields(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Field> values) {
-      if (fieldsBuilder_ == null) {
-        ensureFieldsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, fields_);
-        onChanged();
-      } else {
-        fieldsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder clearFields() {
-      if (fieldsBuilder_ == null) {
-        fields_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000002);
-        onChanged();
-      } else {
-        fieldsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public Builder removeFields(int index) {
-      if (fieldsBuilder_ == null) {
-        ensureFieldsIsMutable();
-        fields_.remove(index);
-        onChanged();
-      } else {
-        fieldsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder getFieldsBuilder(
-        int index) {
-      return getFieldsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder getFieldsOrBuilder(
-        int index) {
-      if (fieldsBuilder_ == null) {
-        return fields_.get(index);  } else {
-        return fieldsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder> 
-         getFieldsOrBuilderList() {
-      if (fieldsBuilder_ != null) {
-        return fieldsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(fields_);
-      }
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder addFieldsBuilder() {
-      return getFieldsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder addFieldsBuilder(
-        int index) {
-      return getFieldsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * The list of fields.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Field fields = 2;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder> 
-         getFieldsBuilderList() {
-      return getFieldsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Field, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder> 
-        getFieldsFieldBuilder() {
-      if (fieldsBuilder_ == null) {
-        fieldsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Field, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder>(
-                fields_,
-                ((bitField0_ & 0x00000002) == 0x00000002),
-                getParentForChildren(),
-                isClean());
-        fields_ = null;
-      }
-      return fieldsBuilder_;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList oneofs_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-    private void ensureOneofsIsMutable() {
-      if (!((bitField0_ & 0x00000004) == 0x00000004)) {
-        oneofs_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList(oneofs_);
-        bitField0_ |= 0x00000004;
-       }
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
-        getOneofsList() {
-      return oneofs_.getUnmodifiableView();
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public int getOneofsCount() {
-      return oneofs_.size();
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public java.lang.String getOneofs(int index) {
-      return oneofs_.get(index);
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getOneofsBytes(int index) {
-      return oneofs_.getByteString(index);
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public Builder setOneofs(
-        int index, java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureOneofsIsMutable();
-      oneofs_.set(index, value);
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public Builder addOneofs(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureOneofsIsMutable();
-      oneofs_.add(value);
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public Builder addAllOneofs(
-        java.lang.Iterable<java.lang.String> values) {
-      ensureOneofsIsMutable();
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-          values, oneofs_);
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public Builder clearOneofs() {
-      oneofs_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-      bitField0_ = (bitField0_ & ~0x00000004);
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The list of types appearing in `oneof` definitions in this type.
-     * </pre>
-     *
-     * <code>repeated string oneofs = 3;</code>
-     */
-    public Builder addOneofsBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      ensureOneofsIsMutable();
-      oneofs_.add(value);
-      onChanged();
-      return this;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_ =
-      java.util.Collections.emptyList();
-    private void ensureOptionsIsMutable() {
-      if (!((bitField0_ & 0x00000008) == 0x00000008)) {
-        options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>(options_);
-        bitField0_ |= 0x00000008;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> optionsBuilder_;
-
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-      if (optionsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(options_);
-      } else {
-        return optionsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public int getOptionsCount() {
-      if (optionsBuilder_ == null) {
-        return options_.size();
-      } else {
-        return optionsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);
-      } else {
-        return optionsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.set(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder addOptions(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder addOptions(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder addAllOptions(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> values) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, options_);
-        onChanged();
-      } else {
-        optionsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder clearOptions() {
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000008);
-        onChanged();
-      } else {
-        optionsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public Builder removeOptions(int index) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.remove(index);
-        onChanged();
-      } else {
-        optionsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder getOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-        int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);  } else {
-        return optionsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-         getOptionsOrBuilderList() {
-      if (optionsBuilder_ != null) {
-        return optionsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(options_);
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder() {
-      return getOptionsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 4;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder> 
-         getOptionsBuilderList() {
-      return getOptionsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-        getOptionsFieldBuilder() {
-      if (optionsBuilder_ == null) {
-        optionsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder>(
-                options_,
-                ((bitField0_ & 0x00000008) == 0x00000008),
-                getParentForChildren(),
-                isClean());
-        options_ = null;
-      }
-      return optionsBuilder_;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext sourceContext_ = null;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder> sourceContextBuilder_;
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public boolean hasSourceContext() {
-      return sourceContextBuilder_ != null || sourceContext_ != null;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext() {
-      if (sourceContextBuilder_ == null) {
-        return sourceContext_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-      } else {
-        return sourceContextBuilder_.getMessage();
-      }
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public Builder setSourceContext(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext value) {
-      if (sourceContextBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        sourceContext_ = value;
-        onChanged();
-      } else {
-        sourceContextBuilder_.setMessage(value);
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public Builder setSourceContext(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder builderForValue) {
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = builderForValue.build();
-        onChanged();
-      } else {
-        sourceContextBuilder_.setMessage(builderForValue.build());
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public Builder mergeSourceContext(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext value) {
-      if (sourceContextBuilder_ == null) {
-        if (sourceContext_ != null) {
-          sourceContext_ =
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.newBuilder(sourceContext_).mergeFrom(value).buildPartial();
-        } else {
-          sourceContext_ = value;
-        }
-        onChanged();
-      } else {
-        sourceContextBuilder_.mergeFrom(value);
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public Builder clearSourceContext() {
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = null;
-        onChanged();
-      } else {
-        sourceContext_ = null;
-        sourceContextBuilder_ = null;
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder getSourceContextBuilder() {
-      
-      onChanged();
-      return getSourceContextFieldBuilder().getBuilder();
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder() {
-      if (sourceContextBuilder_ != null) {
-        return sourceContextBuilder_.getMessageOrBuilder();
-      } else {
-        return sourceContext_ == null ?
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-      }
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder> 
-        getSourceContextFieldBuilder() {
-      if (sourceContextBuilder_ == null) {
-        sourceContextBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder>(
-                getSourceContext(),
-                getParentForChildren(),
-                isClean());
-        sourceContext_ = null;
-      }
-      return sourceContextBuilder_;
-    }
-
-    private int syntax_ = 0;
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 6;</code>
-     */
-    public int getSyntaxValue() {
-      return syntax_;
-    }
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 6;</code>
-     */
-    public Builder setSyntaxValue(int value) {
-      syntax_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.valueOf(syntax_);
-      return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.UNRECOGNIZED : result;
-    }
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 6;</code>
-     */
-    public Builder setSyntax(org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax value) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-      
-      syntax_ = value.getNumber();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 6;</code>
-     */
-    public Builder clearSyntax() {
-      
-      syntax_ = 0;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Type)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Type)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Type DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Type();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Type getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Type>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Type>() {
-    public Type parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Type(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Type> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Type> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Type getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TypeOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TypeOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TypeOrBuilder.java
deleted file mode 100644
index fe845dd..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TypeOrBuilder.java
+++ /dev/null
@@ -1,192 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface TypeOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Type)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The fully qualified message name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  java.lang.String getName();
-  /**
-   * <pre>
-   * The fully qualified message name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes();
-
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Field> 
-      getFieldsList();
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Field getFields(int index);
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  int getFieldsCount();
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder> 
-      getFieldsOrBuilderList();
-  /**
-   * <pre>
-   * The list of fields.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Field fields = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder getFieldsOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * The list of types appearing in `oneof` definitions in this type.
-   * </pre>
-   *
-   * <code>repeated string oneofs = 3;</code>
-   */
-  java.util.List<java.lang.String>
-      getOneofsList();
-  /**
-   * <pre>
-   * The list of types appearing in `oneof` definitions in this type.
-   * </pre>
-   *
-   * <code>repeated string oneofs = 3;</code>
-   */
-  int getOneofsCount();
-  /**
-   * <pre>
-   * The list of types appearing in `oneof` definitions in this type.
-   * </pre>
-   *
-   * <code>repeated string oneofs = 3;</code>
-   */
-  java.lang.String getOneofs(int index);
-  /**
-   * <pre>
-   * The list of types appearing in `oneof` definitions in this type.
-   * </pre>
-   *
-   * <code>repeated string oneofs = 3;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getOneofsBytes(int index);
-
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> 
-      getOptionsList();
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index);
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  int getOptionsCount();
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList();
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 4;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  boolean hasSourceContext();
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext();
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder();
-
-  /**
-   * <pre>
-   * The source syntax.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 6;</code>
-   */
-  int getSyntaxValue();
-  /**
-   * <pre>
-   * The source syntax.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 6;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TypeProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TypeProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TypeProto.java
deleted file mode 100644
index e7c3094..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TypeProto.java
+++ /dev/null
@@ -1,141 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class TypeProto {
-  private TypeProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Type_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Type_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Field_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Field_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Enum_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Enum_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_EnumValue_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_EnumValue_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Option_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Option_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\032google/protobuf/type.proto\022\017google.pro" +
-      "tobuf\032\031google/protobuf/any.proto\032$google" +
-      "/protobuf/source_context.proto\"\327\001\n\004Type\022" +
-      "\014\n\004name\030\001 \001(\t\022&\n\006fields\030\002 \003(\0132\026.google.p" +
-      "rotobuf.Field\022\016\n\006oneofs\030\003 \003(\t\022(\n\007options" +
-      "\030\004 \003(\0132\027.google.protobuf.Option\0226\n\016sourc" +
-      "e_context\030\005 \001(\0132\036.google.protobuf.Source" +
-      "Context\022\'\n\006syntax\030\006 \001(\0162\027.google.protobu" +
-      "f.Syntax\"\325\005\n\005Field\022)\n\004kind\030\001 \001(\0162\033.googl" +
-      "e.protobuf.Field.Kind\0227\n\013cardinality\030\002 \001",
-      "(\0162\".google.protobuf.Field.Cardinality\022\016" +
-      "\n\006number\030\003 \001(\005\022\014\n\004name\030\004 \001(\t\022\020\n\010type_url" +
-      "\030\006 \001(\t\022\023\n\013oneof_index\030\007 \001(\005\022\016\n\006packed\030\010 " +
-      "\001(\010\022(\n\007options\030\t \003(\0132\027.google.protobuf.O" +
-      "ption\022\021\n\tjson_name\030\n \001(\t\022\025\n\rdefault_valu" +
-      "e\030\013 \001(\t\"\310\002\n\004Kind\022\020\n\014TYPE_UNKNOWN\020\000\022\017\n\013TY" +
-      "PE_DOUBLE\020\001\022\016\n\nTYPE_FLOAT\020\002\022\016\n\nTYPE_INT6" +
-      "4\020\003\022\017\n\013TYPE_UINT64\020\004\022\016\n\nTYPE_INT32\020\005\022\020\n\014" +
-      "TYPE_FIXED64\020\006\022\020\n\014TYPE_FIXED32\020\007\022\r\n\tTYPE" +
-      "_BOOL\020\010\022\017\n\013TYPE_STRING\020\t\022\016\n\nTYPE_GROUP\020\n",
-      "\022\020\n\014TYPE_MESSAGE\020\013\022\016\n\nTYPE_BYTES\020\014\022\017\n\013TY" +
-      "PE_UINT32\020\r\022\r\n\tTYPE_ENUM\020\016\022\021\n\rTYPE_SFIXE" +
-      "D32\020\017\022\021\n\rTYPE_SFIXED64\020\020\022\017\n\013TYPE_SINT32\020" +
-      "\021\022\017\n\013TYPE_SINT64\020\022\"t\n\013Cardinality\022\027\n\023CAR" +
-      "DINALITY_UNKNOWN\020\000\022\030\n\024CARDINALITY_OPTION" +
-      "AL\020\001\022\030\n\024CARDINALITY_REQUIRED\020\002\022\030\n\024CARDIN" +
-      "ALITY_REPEATED\020\003\"\316\001\n\004Enum\022\014\n\004name\030\001 \001(\t\022" +
-      "-\n\tenumvalue\030\002 \003(\0132\032.google.protobuf.Enu" +
-      "mValue\022(\n\007options\030\003 \003(\0132\027.google.protobu" +
-      "f.Option\0226\n\016source_context\030\004 \001(\0132\036.googl",
-      "e.protobuf.SourceContext\022\'\n\006syntax\030\005 \001(\016" +
-      "2\027.google.protobuf.Syntax\"S\n\tEnumValue\022\014" +
-      "\n\004name\030\001 \001(\t\022\016\n\006number\030\002 \001(\005\022(\n\007options\030" +
-      "\003 \003(\0132\027.google.protobuf.Option\";\n\006Option" +
-      "\022\014\n\004name\030\001 \001(\t\022#\n\005value\030\002 \001(\0132\024.google.p" +
-      "rotobuf.Any*.\n\006Syntax\022\021\n\rSYNTAX_PROTO2\020\000" +
-      "\022\021\n\rSYNTAX_PROTO3\020\001B}\n\023com.google.protob" +
-      "ufB\tTypeProtoP\001Z/google.golang.org/genpr" +
-      "oto/protobuf/ptype;ptype\370\001\001\242\002\003GPB\252\002\036Goog" +
-      "le.Protobuf.WellKnownTypesb\006proto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyProto.getDescriptor(),
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.getDescriptor(),
-        }, assigner);
-    internal_static_google_protobuf_Type_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_Type_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Type_descriptor,
-        new java.lang.String[] { "Name", "Fields", "Oneofs", "Options", "SourceContext", "Syntax", });
-    internal_static_google_protobuf_Field_descriptor =
-      getDescriptor().getMessageTypes().get(1);
-    internal_static_google_protobuf_Field_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Field_descriptor,
-        new java.lang.String[] { "Kind", "Cardinality", "Number", "Name", "TypeUrl", "OneofIndex", "Packed", "Options", "JsonName", "DefaultValue", });
-    internal_static_google_protobuf_Enum_descriptor =
-      getDescriptor().getMessageTypes().get(2);
-    internal_static_google_protobuf_Enum_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Enum_descriptor,
-        new java.lang.String[] { "Name", "Enumvalue", "Options", "SourceContext", "Syntax", });
-    internal_static_google_protobuf_EnumValue_descriptor =
-      getDescriptor().getMessageTypes().get(3);
-    internal_static_google_protobuf_EnumValue_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_EnumValue_descriptor,
-        new java.lang.String[] { "Name", "Number", "Options", });
-    internal_static_google_protobuf_Option_descriptor =
-      getDescriptor().getMessageTypes().get(4);
-    internal_static_google_protobuf_Option_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Option_descriptor,
-        new java.lang.String[] { "Name", "Value", });
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyProto.getDescriptor();
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.getDescriptor();
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[17/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RepeatedFieldBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RepeatedFieldBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RepeatedFieldBuilder.java
deleted file mode 100644
index 8968a70..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RepeatedFieldBuilder.java
+++ /dev/null
@@ -1,708 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.AbstractList;
-import java.util.ArrayList;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.List;
-
-/**
- * {@code RepeatedFieldBuilder} implements a structure that a protocol
- * message uses to hold a repeated field of other protocol messages. It supports
- * the classical use case of adding immutable {@link Message}'s to the
- * repeated field and is highly optimized around this (no extra memory
- * allocations and sharing of immutable arrays).
- * <br>
- * It also supports the additional use case of adding a {@link Message.Builder}
- * to the repeated field and deferring conversion of that {@code Builder}
- * to an immutable {@code Message}. In this way, it's possible to maintain
- * a tree of {@code Builder}'s that acts as a fully read/write data
- * structure.
- * <br>
- * Logically, one can think of a tree of builders as converting the entire tree
- * to messages when build is called on the root or when any method is called
- * that desires a Message instead of a Builder. In terms of the implementation,
- * the {@code SingleFieldBuilder} and {@code RepeatedFieldBuilder}
- * classes cache messages that were created so that messages only need to be
- * created when some change occurred in its builder or a builder for one of its
- * descendants.
- *
- * @param <MType> the type of message for the field
- * @param <BType> the type of builder for the field
- * @param <IType> the common interface for the message and the builder
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public class RepeatedFieldBuilder
-    <MType extends GeneratedMessage,
-     BType extends GeneratedMessage.Builder,
-     IType extends MessageOrBuilder>
-    implements GeneratedMessage.BuilderParent {
-
-  // Parent to send changes to.
-  private GeneratedMessage.BuilderParent parent;
-
-  // List of messages. Never null. It may be immutable, in which case
-  // isMessagesListMutable will be false. See note below.
-  private List<MType> messages;
-
-  // Whether messages is an mutable array that can be modified.
-  private boolean isMessagesListMutable;
-
-  // List of builders. May be null, in which case, no nested builders were
-  // created. If not null, entries represent the builder for that index.
-  private List<SingleFieldBuilder<MType, BType, IType>> builders;
-
-  // Here are the invariants for messages and builders:
-  // 1. messages is never null and its count corresponds to the number of items
-  //    in the repeated field.
-  // 2. If builders is non-null, messages and builders MUST always
-  //    contain the same number of items.
-  // 3. Entries in either array can be null, but for any index, there MUST be
-  //    either a Message in messages or a builder in builders.
-  // 4. If the builder at an index is non-null, the builder is
-  //    authoritative. This is the case where a Builder was set on the index.
-  //    Any message in the messages array MUST be ignored.
-  // t. If the builder at an index is null, the message in the messages
-  //    list is authoritative. This is the case where a Message (not a Builder)
-  //    was set directly for an index.
-
-  // Indicates that we've built a message and so we are now obligated
-  // to dispatch dirty invalidations. See GeneratedMessage.BuilderListener.
-  private boolean isClean;
-
-  // A view of this builder that exposes a List interface of messages. This is
-  // initialized on demand. This is fully backed by this object and all changes
-  // are reflected in it. Access to any item converts it to a message if it
-  // was a builder.
-  private MessageExternalList<MType, BType, IType> externalMessageList;
-
-  // A view of this builder that exposes a List interface of builders. This is
-  // initialized on demand. This is fully backed by this object and all changes
-  // are reflected in it. Access to any item converts it to a builder if it
-  // was a message.
-  private BuilderExternalList<MType, BType, IType> externalBuilderList;
-
-  // A view of this builder that exposes a List interface of the interface
-  // implemented by messages and builders. This is initialized on demand. This
-  // is fully backed by this object and all changes are reflected in it.
-  // Access to any item returns either a builder or message depending on
-  // what is most efficient.
-  private MessageOrBuilderExternalList<MType, BType, IType>
-      externalMessageOrBuilderList;
-
-  /**
-   * Constructs a new builder with an empty list of messages.
-   *
-   * @param messages the current list of messages
-   * @param isMessagesListMutable Whether the messages list is mutable
-   * @param parent a listener to notify of changes
-   * @param isClean whether the builder is initially marked clean
-   */
-  public RepeatedFieldBuilder(
-      List<MType> messages,
-      boolean isMessagesListMutable,
-      GeneratedMessage.BuilderParent parent,
-      boolean isClean) {
-    this.messages = messages;
-    this.isMessagesListMutable = isMessagesListMutable;
-    this.parent = parent;
-    this.isClean = isClean;
-  }
-
-  public void dispose() {
-    // Null out parent so we stop sending it invalidations.
-    parent = null;
-  }
-
-  /**
-   * Ensures that the list of messages is mutable so it can be updated. If it's
-   * immutable, a copy is made.
-   */
-  private void ensureMutableMessageList() {
-    if (!isMessagesListMutable) {
-      messages = new ArrayList<MType>(messages);
-      isMessagesListMutable = true;
-    }
-  }
-
-  /**
-   * Ensures that the list of builders is not null. If it's null, the list is
-   * created and initialized to be the same size as the messages list with
-   * null entries.
-   */
-  private void ensureBuilders() {
-    if (this.builders == null) {
-      this.builders =
-          new ArrayList<SingleFieldBuilder<MType, BType, IType>>(
-              messages.size());
-      for (int i = 0; i < messages.size(); i++) {
-        builders.add(null);
-      }
-    }
-  }
-
-  /**
-   * Gets the count of items in the list.
-   *
-   * @return the count of items in the list.
-   */
-  public int getCount() {
-    return messages.size();
-  }
-
-  /**
-   * Gets whether the list is empty.
-   *
-   * @return whether the list is empty
-   */
-  public boolean isEmpty() {
-    return messages.isEmpty();
-  }
-
-  /**
-   * Get the message at the specified index. If the message is currently stored
-   * as a {@code Builder}, it is converted to a {@code Message} by
-   * calling {@link Message.Builder#buildPartial} on it.
-   *
-   * @param index the index of the message to get
-   * @return the message for the specified index
-   */
-  public MType getMessage(int index) {
-    return getMessage(index, false);
-  }
-
-  /**
-   * Get the message at the specified index. If the message is currently stored
-   * as a {@code Builder}, it is converted to a {@code Message} by
-   * calling {@link Message.Builder#buildPartial} on it.
-   *
-   * @param index the index of the message to get
-   * @param forBuild this is being called for build so we want to make sure
-   *     we SingleFieldBuilder.build to send dirty invalidations
-   * @return the message for the specified index
-   */
-  private MType getMessage(int index, boolean forBuild) {
-    if (this.builders == null) {
-      // We don't have any builders -- return the current Message.
-      // This is the case where no builder was created, so we MUST have a
-      // Message.
-      return messages.get(index);
-    }
-
-    SingleFieldBuilder<MType, BType, IType> builder = builders.get(index);
-    if (builder == null) {
-      // We don't have a builder -- return the current message.
-      // This is the case where no builder was created for the entry at index,
-      // so we MUST have a message.
-      return messages.get(index);
-
-    } else {
-      return forBuild ? builder.build() : builder.getMessage();
-    }
-  }
-
-  /**
-   * Gets a builder for the specified index. If no builder has been created for
-   * that index, a builder is created on demand by calling
-   * {@link Message#toBuilder}.
-   *
-   * @param index the index of the message to get
-   * @return The builder for that index
-   */
-  public BType getBuilder(int index) {
-    ensureBuilders();
-    SingleFieldBuilder<MType, BType, IType> builder = builders.get(index);
-    if (builder == null) {
-      MType message = messages.get(index);
-      builder = new SingleFieldBuilder<MType, BType, IType>(
-          message, this, isClean);
-      builders.set(index, builder);
-    }
-    return builder.getBuilder();
-  }
-
-  /**
-   * Gets the base class interface for the specified index. This may either be
-   * a builder or a message. It will return whatever is more efficient.
-   *
-   * @param index the index of the message to get
-   * @return the message or builder for the index as the base class interface
-   */
-  @SuppressWarnings("unchecked")
-  public IType getMessageOrBuilder(int index) {
-    if (this.builders == null) {
-      // We don't have any builders -- return the current Message.
-      // This is the case where no builder was created, so we MUST have a
-      // Message.
-      return (IType) messages.get(index);
-    }
-
-    SingleFieldBuilder<MType, BType, IType> builder = builders.get(index);
-    if (builder == null) {
-      // We don't have a builder -- return the current message.
-      // This is the case where no builder was created for the entry at index,
-      // so we MUST have a message.
-      return (IType) messages.get(index);
-
-    } else {
-      return builder.getMessageOrBuilder();
-    }
-  }
-
-  /**
-   * Sets a  message at the specified index replacing the existing item at
-   * that index.
-   *
-   * @param index the index to set.
-   * @param message the message to set
-   * @return the builder
-   */
-  public RepeatedFieldBuilder<MType, BType, IType> setMessage(
-      int index, MType message) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    ensureMutableMessageList();
-    messages.set(index, message);
-    if (builders != null) {
-      SingleFieldBuilder<MType, BType, IType> entry =
-          builders.set(index, null);
-      if (entry != null) {
-        entry.dispose();
-      }
-    }
-    onChanged();
-    incrementModCounts();
-    return this;
-  }
-
-  /**
-   * Appends the specified element to the end of this list.
-   *
-   * @param message the message to add
-   * @return the builder
-   */
-  public RepeatedFieldBuilder<MType, BType, IType> addMessage(
-      MType message) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    ensureMutableMessageList();
-    messages.add(message);
-    if (builders != null) {
-      builders.add(null);
-    }
-    onChanged();
-    incrementModCounts();
-    return this;
-  }
-
-  /**
-   * Inserts the specified message at the specified position in this list.
-   * Shifts the element currently at that position (if any) and any subsequent
-   * elements to the right (adds one to their indices).
-   *
-   * @param index the index at which to insert the message
-   * @param message the message to add
-   * @return the builder
-   */
-  public RepeatedFieldBuilder<MType, BType, IType> addMessage(
-      int index, MType message) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    ensureMutableMessageList();
-    messages.add(index, message);
-    if (builders != null) {
-      builders.add(index, null);
-    }
-    onChanged();
-    incrementModCounts();
-    return this;
-  }
-
-  /**
-   * Appends all of the messages in the specified collection to the end of
-   * this list, in the order that they are returned by the specified
-   * collection's iterator.
-   *
-   * @param values the messages to add
-   * @return the builder
-   */
-  public RepeatedFieldBuilder<MType, BType, IType> addAllMessages(
-      Iterable<? extends MType> values) {
-    for (final MType value : values) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-    }
-
-    // If we can inspect the size, we can more efficiently add messages.
-    int size = -1;
-    if (values instanceof Collection) {
-      @SuppressWarnings("unchecked") final
-      Collection<MType> collection = (Collection<MType>) values;
-      if (collection.size() == 0) {
-        return this;
-      }
-      size = collection.size();
-    }
-    ensureMutableMessageList();
-
-    if (size >= 0 && messages instanceof ArrayList) {
-      ((ArrayList<MType>) messages)
-          .ensureCapacity(messages.size() + size);
-    }
-
-    for (MType value : values) {
-      addMessage(value);
-    }
-
-    onChanged();
-    incrementModCounts();
-    return this;
-  }
-
-  /**
-   * Appends a new builder to the end of this list and returns the builder.
-   *
-   * @param message the message to add which is the basis of the builder
-   * @return the new builder
-   */
-  public BType addBuilder(MType message) {
-    ensureMutableMessageList();
-    ensureBuilders();
-    SingleFieldBuilder<MType, BType, IType> builder =
-        new SingleFieldBuilder<MType, BType, IType>(
-            message, this, isClean);
-    messages.add(null);
-    builders.add(builder);
-    onChanged();
-    incrementModCounts();
-    return builder.getBuilder();
-  }
-
-  /**
-   * Inserts a new builder at the specified position in this list.
-   * Shifts the element currently at that position (if any) and any subsequent
-   * elements to the right (adds one to their indices).
-   *
-   * @param index the index at which to insert the builder
-   * @param message the message to add which is the basis of the builder
-   * @return the builder
-   */
-  public BType addBuilder(int index, MType message) {
-    ensureMutableMessageList();
-    ensureBuilders();
-    SingleFieldBuilder<MType, BType, IType> builder =
-        new SingleFieldBuilder<MType, BType, IType>(
-            message, this, isClean);
-    messages.add(index, null);
-    builders.add(index, builder);
-    onChanged();
-    incrementModCounts();
-    return builder.getBuilder();
-  }
-
-  /**
-   * Removes the element at the specified position in this list. Shifts any
-   * subsequent elements to the left (subtracts one from their indices).
-   * Returns the element that was removed from the list.
-   *
-   * @param index the index at which to remove the message
-   */
-  public void remove(int index) {
-    ensureMutableMessageList();
-    messages.remove(index);
-    if (builders != null) {
-      SingleFieldBuilder<MType, BType, IType> entry =
-          builders.remove(index);
-      if (entry != null) {
-        entry.dispose();
-      }
-    }
-    onChanged();
-    incrementModCounts();
-  }
-
-  /**
-   * Removes all of the elements from this list.
-   * The list will be empty after this call returns.
-   */
-  public void clear() {
-    messages = Collections.emptyList();
-    isMessagesListMutable = false;
-    if (builders != null) {
-      for (SingleFieldBuilder<MType, BType, IType> entry :
-          builders) {
-        if (entry != null) {
-          entry.dispose();
-        }
-      }
-      builders = null;
-    }
-    onChanged();
-    incrementModCounts();
-  }
-
-  /**
-   * Builds the list of messages from the builder and returns them.
-   *
-   * @return an immutable list of messages
-   */
-  public List<MType> build() {
-    // Now that build has been called, we are required to dispatch
-    // invalidations.
-    isClean = true;
-
-    if (!isMessagesListMutable && builders == null) {
-      // We still have an immutable list and we never created a builder.
-      return messages;
-    }
-
-    boolean allMessagesInSync = true;
-    if (!isMessagesListMutable) {
-      // We still have an immutable list. Let's see if any of them are out
-      // of sync with their builders.
-      for (int i = 0; i < messages.size(); i++) {
-        Message message = messages.get(i);
-        SingleFieldBuilder<MType, BType, IType> builder = builders.get(i);
-        if (builder != null) {
-          if (builder.build() != message) {
-            allMessagesInSync = false;
-            break;
-          }
-        }
-      }
-      if (allMessagesInSync) {
-        // Immutable list is still in sync.
-        return messages;
-      }
-    }
-
-    // Need to make sure messages is up to date
-    ensureMutableMessageList();
-    for (int i = 0; i < messages.size(); i++) {
-      messages.set(i, getMessage(i, true));
-    }
-
-    // We're going to return our list as immutable so we mark that we can
-    // no longer update it.
-    messages = Collections.unmodifiableList(messages);
-    isMessagesListMutable = false;
-    return messages;
-  }
-
-  /**
-   * Gets a view of the builder as a list of messages. The returned list is live
-   * and will reflect any changes to the underlying builder.
-   *
-   * @return the messages in the list
-   */
-  public List<MType> getMessageList() {
-    if (externalMessageList == null) {
-      externalMessageList =
-          new MessageExternalList<MType, BType, IType>(this);
-    }
-    return externalMessageList;
-  }
-
-  /**
-   * Gets a view of the builder as a list of builders. This returned list is
-   * live and will reflect any changes to the underlying builder.
-   *
-   * @return the builders in the list
-   */
-  public List<BType> getBuilderList() {
-    if (externalBuilderList == null) {
-      externalBuilderList =
-          new BuilderExternalList<MType, BType, IType>(this);
-    }
-    return externalBuilderList;
-  }
-
-  /**
-   * Gets a view of the builder as a list of MessageOrBuilders. This returned
-   * list is live and will reflect any changes to the underlying builder.
-   *
-   * @return the builders in the list
-   */
-  public List<IType> getMessageOrBuilderList() {
-    if (externalMessageOrBuilderList == null) {
-      externalMessageOrBuilderList =
-          new MessageOrBuilderExternalList<MType, BType, IType>(this);
-    }
-    return externalMessageOrBuilderList;
-  }
-
-  /**
-   * Called when a the builder or one of its nested children has changed
-   * and any parent should be notified of its invalidation.
-   */
-  private void onChanged() {
-    if (isClean && parent != null) {
-      parent.markDirty();
-
-      // Don't keep dispatching invalidations until build is called again.
-      isClean = false;
-    }
-  }
-
-  @Override
-  public void markDirty() {
-    onChanged();
-  }
-
-  /**
-   * Increments the mod counts so that an ConcurrentModificationException can
-   * be thrown if calling code tries to modify the builder while its iterating
-   * the list.
-   */
-  private void incrementModCounts() {
-    if (externalMessageList != null) {
-      externalMessageList.incrementModCount();
-    }
-    if (externalBuilderList != null) {
-      externalBuilderList.incrementModCount();
-    }
-    if (externalMessageOrBuilderList != null) {
-      externalMessageOrBuilderList.incrementModCount();
-    }
-  }
-
-  /**
-   * Provides a live view of the builder as a list of messages.
-   *
-   * @param <MType> the type of message for the field
-   * @param <BType> the type of builder for the field
-   * @param <IType> the common interface for the message and the builder
-   */
-  private static class MessageExternalList<
-      MType extends GeneratedMessage,
-      BType extends GeneratedMessage.Builder,
-      IType extends MessageOrBuilder>
-      extends AbstractList<MType> implements List<MType> {
-
-    RepeatedFieldBuilder<MType, BType, IType> builder;
-
-    MessageExternalList(
-        RepeatedFieldBuilder<MType, BType, IType> builder) {
-      this.builder = builder;
-    }
-
-    @Override
-    public int size() {
-      return this.builder.getCount();
-    }
-
-    @Override
-    public MType get(int index) {
-      return builder.getMessage(index);
-    }
-
-    void incrementModCount() {
-      modCount++;
-    }
-  }
-
-  /**
-   * Provides a live view of the builder as a list of builders.
-   *
-   * @param <MType> the type of message for the field
-   * @param <BType> the type of builder for the field
-   * @param <IType> the common interface for the message and the builder
-   */
-  private static class BuilderExternalList<
-      MType extends GeneratedMessage,
-      BType extends GeneratedMessage.Builder,
-      IType extends MessageOrBuilder>
-      extends AbstractList<BType> implements List<BType> {
-
-    RepeatedFieldBuilder<MType, BType, IType> builder;
-
-    BuilderExternalList(
-        RepeatedFieldBuilder<MType, BType, IType> builder) {
-      this.builder = builder;
-    }
-
-    @Override
-    public int size() {
-      return this.builder.getCount();
-    }
-
-    @Override
-    public BType get(int index) {
-      return builder.getBuilder(index);
-    }
-
-    void incrementModCount() {
-      modCount++;
-    }
-  }
-
-  /**
-   * Provides a live view of the builder as a list of builders.
-   *
-   * @param <MType> the type of message for the field
-   * @param <BType> the type of builder for the field
-   * @param <IType> the common interface for the message and the builder
-   */
-  private static class MessageOrBuilderExternalList<
-      MType extends GeneratedMessage,
-      BType extends GeneratedMessage.Builder,
-      IType extends MessageOrBuilder>
-      extends AbstractList<IType> implements List<IType> {
-
-    RepeatedFieldBuilder<MType, BType, IType> builder;
-
-    MessageOrBuilderExternalList(
-        RepeatedFieldBuilder<MType, BType, IType> builder) {
-      this.builder = builder;
-    }
-
-    @Override
-    public int size() {
-      return this.builder.getCount();
-    }
-
-    @Override
-    public IType get(int index) {
-      return builder.getMessageOrBuilder(index);
-    }
-
-    void incrementModCount() {
-      modCount++;
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RepeatedFieldBuilderV3.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RepeatedFieldBuilderV3.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RepeatedFieldBuilderV3.java
deleted file mode 100644
index 9963e05..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RepeatedFieldBuilderV3.java
+++ /dev/null
@@ -1,708 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.AbstractList;
-import java.util.ArrayList;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.List;
-
-/**
- * {@code RepeatedFieldBuilderV3} implements a structure that a protocol
- * message uses to hold a repeated field of other protocol messages. It supports
- * the classical use case of adding immutable {@link Message}'s to the
- * repeated field and is highly optimized around this (no extra memory
- * allocations and sharing of immutable arrays).
- * <br>
- * It also supports the additional use case of adding a {@link Message.Builder}
- * to the repeated field and deferring conversion of that {@code Builder}
- * to an immutable {@code Message}. In this way, it's possible to maintain
- * a tree of {@code Builder}'s that acts as a fully read/write data
- * structure.
- * <br>
- * Logically, one can think of a tree of builders as converting the entire tree
- * to messages when build is called on the root or when any method is called
- * that desires a Message instead of a Builder. In terms of the implementation,
- * the {@code SingleFieldBuilderV3} and {@code RepeatedFieldBuilderV3}
- * classes cache messages that were created so that messages only need to be
- * created when some change occurred in its builder or a builder for one of its
- * descendants.
- *
- * @param <MType> the type of message for the field
- * @param <BType> the type of builder for the field
- * @param <IType> the common interface for the message and the builder
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public class RepeatedFieldBuilderV3
-    <MType extends AbstractMessage,
-     BType extends AbstractMessage.Builder,
-     IType extends MessageOrBuilder>
-    implements AbstractMessage.BuilderParent {
-
-  // Parent to send changes to.
-  private AbstractMessage.BuilderParent parent;
-
-  // List of messages. Never null. It may be immutable, in which case
-  // isMessagesListMutable will be false. See note below.
-  private List<MType> messages;
-
-  // Whether messages is an mutable array that can be modified.
-  private boolean isMessagesListMutable;
-
-  // List of builders. May be null, in which case, no nested builders were
-  // created. If not null, entries represent the builder for that index.
-  private List<SingleFieldBuilderV3<MType, BType, IType>> builders;
-
-  // Here are the invariants for messages and builders:
-  // 1. messages is never null and its count corresponds to the number of items
-  //    in the repeated field.
-  // 2. If builders is non-null, messages and builders MUST always
-  //    contain the same number of items.
-  // 3. Entries in either array can be null, but for any index, there MUST be
-  //    either a Message in messages or a builder in builders.
-  // 4. If the builder at an index is non-null, the builder is
-  //    authoritative. This is the case where a Builder was set on the index.
-  //    Any message in the messages array MUST be ignored.
-  // t. If the builder at an index is null, the message in the messages
-  //    list is authoritative. This is the case where a Message (not a Builder)
-  //    was set directly for an index.
-
-  // Indicates that we've built a message and so we are now obligated
-  // to dispatch dirty invalidations. See AbstractMessage.BuilderListener.
-  private boolean isClean;
-
-  // A view of this builder that exposes a List interface of messages. This is
-  // initialized on demand. This is fully backed by this object and all changes
-  // are reflected in it. Access to any item converts it to a message if it
-  // was a builder.
-  private MessageExternalList<MType, BType, IType> externalMessageList;
-
-  // A view of this builder that exposes a List interface of builders. This is
-  // initialized on demand. This is fully backed by this object and all changes
-  // are reflected in it. Access to any item converts it to a builder if it
-  // was a message.
-  private BuilderExternalList<MType, BType, IType> externalBuilderList;
-
-  // A view of this builder that exposes a List interface of the interface
-  // implemented by messages and builders. This is initialized on demand. This
-  // is fully backed by this object and all changes are reflected in it.
-  // Access to any item returns either a builder or message depending on
-  // what is most efficient.
-  private MessageOrBuilderExternalList<MType, BType, IType>
-      externalMessageOrBuilderList;
-
-  /**
-   * Constructs a new builder with an empty list of messages.
-   *
-   * @param messages the current list of messages
-   * @param isMessagesListMutable Whether the messages list is mutable
-   * @param parent a listener to notify of changes
-   * @param isClean whether the builder is initially marked clean
-   */
-  public RepeatedFieldBuilderV3(
-      List<MType> messages,
-      boolean isMessagesListMutable,
-      AbstractMessage.BuilderParent parent,
-      boolean isClean) {
-    this.messages = messages;
-    this.isMessagesListMutable = isMessagesListMutable;
-    this.parent = parent;
-    this.isClean = isClean;
-  }
-
-  public void dispose() {
-    // Null out parent so we stop sending it invalidations.
-    parent = null;
-  }
-
-  /**
-   * Ensures that the list of messages is mutable so it can be updated. If it's
-   * immutable, a copy is made.
-   */
-  private void ensureMutableMessageList() {
-    if (!isMessagesListMutable) {
-      messages = new ArrayList<MType>(messages);
-      isMessagesListMutable = true;
-    }
-  }
-
-  /**
-   * Ensures that the list of builders is not null. If it's null, the list is
-   * created and initialized to be the same size as the messages list with
-   * null entries.
-   */
-  private void ensureBuilders() {
-    if (this.builders == null) {
-      this.builders =
-          new ArrayList<SingleFieldBuilderV3<MType, BType, IType>>(
-              messages.size());
-      for (int i = 0; i < messages.size(); i++) {
-        builders.add(null);
-      }
-    }
-  }
-
-  /**
-   * Gets the count of items in the list.
-   *
-   * @return the count of items in the list.
-   */
-  public int getCount() {
-    return messages.size();
-  }
-
-  /**
-   * Gets whether the list is empty.
-   *
-   * @return whether the list is empty
-   */
-  public boolean isEmpty() {
-    return messages.isEmpty();
-  }
-
-  /**
-   * Get the message at the specified index. If the message is currently stored
-   * as a {@code Builder}, it is converted to a {@code Message} by
-   * calling {@link Message.Builder#buildPartial} on it.
-   *
-   * @param index the index of the message to get
-   * @return the message for the specified index
-   */
-  public MType getMessage(int index) {
-    return getMessage(index, false);
-  }
-
-  /**
-   * Get the message at the specified index. If the message is currently stored
-   * as a {@code Builder}, it is converted to a {@code Message} by
-   * calling {@link Message.Builder#buildPartial} on it.
-   *
-   * @param index the index of the message to get
-   * @param forBuild this is being called for build so we want to make sure
-   *     we SingleFieldBuilderV3.build to send dirty invalidations
-   * @return the message for the specified index
-   */
-  private MType getMessage(int index, boolean forBuild) {
-    if (this.builders == null) {
-      // We don't have any builders -- return the current Message.
-      // This is the case where no builder was created, so we MUST have a
-      // Message.
-      return messages.get(index);
-    }
-
-    SingleFieldBuilderV3<MType, BType, IType> builder = builders.get(index);
-    if (builder == null) {
-      // We don't have a builder -- return the current message.
-      // This is the case where no builder was created for the entry at index,
-      // so we MUST have a message.
-      return messages.get(index);
-
-    } else {
-      return forBuild ? builder.build() : builder.getMessage();
-    }
-  }
-
-  /**
-   * Gets a builder for the specified index. If no builder has been created for
-   * that index, a builder is created on demand by calling
-   * {@link Message#toBuilder}.
-   *
-   * @param index the index of the message to get
-   * @return The builder for that index
-   */
-  public BType getBuilder(int index) {
-    ensureBuilders();
-    SingleFieldBuilderV3<MType, BType, IType> builder = builders.get(index);
-    if (builder == null) {
-      MType message = messages.get(index);
-      builder = new SingleFieldBuilderV3<MType, BType, IType>(
-          message, this, isClean);
-      builders.set(index, builder);
-    }
-    return builder.getBuilder();
-  }
-
-  /**
-   * Gets the base class interface for the specified index. This may either be
-   * a builder or a message. It will return whatever is more efficient.
-   *
-   * @param index the index of the message to get
-   * @return the message or builder for the index as the base class interface
-   */
-  @SuppressWarnings("unchecked")
-  public IType getMessageOrBuilder(int index) {
-    if (this.builders == null) {
-      // We don't have any builders -- return the current Message.
-      // This is the case where no builder was created, so we MUST have a
-      // Message.
-      return (IType) messages.get(index);
-    }
-
-    SingleFieldBuilderV3<MType, BType, IType> builder = builders.get(index);
-    if (builder == null) {
-      // We don't have a builder -- return the current message.
-      // This is the case where no builder was created for the entry at index,
-      // so we MUST have a message.
-      return (IType) messages.get(index);
-
-    } else {
-      return builder.getMessageOrBuilder();
-    }
-  }
-
-  /**
-   * Sets a  message at the specified index replacing the existing item at
-   * that index.
-   *
-   * @param index the index to set.
-   * @param message the message to set
-   * @return the builder
-   */
-  public RepeatedFieldBuilderV3<MType, BType, IType> setMessage(
-      int index, MType message) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    ensureMutableMessageList();
-    messages.set(index, message);
-    if (builders != null) {
-      SingleFieldBuilderV3<MType, BType, IType> entry =
-          builders.set(index, null);
-      if (entry != null) {
-        entry.dispose();
-      }
-    }
-    onChanged();
-    incrementModCounts();
-    return this;
-  }
-
-  /**
-   * Appends the specified element to the end of this list.
-   *
-   * @param message the message to add
-   * @return the builder
-   */
-  public RepeatedFieldBuilderV3<MType, BType, IType> addMessage(
-      MType message) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    ensureMutableMessageList();
-    messages.add(message);
-    if (builders != null) {
-      builders.add(null);
-    }
-    onChanged();
-    incrementModCounts();
-    return this;
-  }
-
-  /**
-   * Inserts the specified message at the specified position in this list.
-   * Shifts the element currently at that position (if any) and any subsequent
-   * elements to the right (adds one to their indices).
-   *
-   * @param index the index at which to insert the message
-   * @param message the message to add
-   * @return the builder
-   */
-  public RepeatedFieldBuilderV3<MType, BType, IType> addMessage(
-      int index, MType message) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    ensureMutableMessageList();
-    messages.add(index, message);
-    if (builders != null) {
-      builders.add(index, null);
-    }
-    onChanged();
-    incrementModCounts();
-    return this;
-  }
-
-  /**
-   * Appends all of the messages in the specified collection to the end of
-   * this list, in the order that they are returned by the specified
-   * collection's iterator.
-   *
-   * @param values the messages to add
-   * @return the builder
-   */
-  public RepeatedFieldBuilderV3<MType, BType, IType> addAllMessages(
-      Iterable<? extends MType> values) {
-    for (final MType value : values) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-    }
-
-    // If we can inspect the size, we can more efficiently add messages.
-    int size = -1;
-    if (values instanceof Collection) {
-      @SuppressWarnings("unchecked") final
-      Collection<MType> collection = (Collection<MType>) values;
-      if (collection.size() == 0) {
-        return this;
-      }
-      size = collection.size();
-    }
-    ensureMutableMessageList();
-
-    if (size >= 0 && messages instanceof ArrayList) {
-      ((ArrayList<MType>) messages)
-          .ensureCapacity(messages.size() + size);
-    }
-
-    for (MType value : values) {
-      addMessage(value);
-    }
-
-    onChanged();
-    incrementModCounts();
-    return this;
-  }
-
-  /**
-   * Appends a new builder to the end of this list and returns the builder.
-   *
-   * @param message the message to add which is the basis of the builder
-   * @return the new builder
-   */
-  public BType addBuilder(MType message) {
-    ensureMutableMessageList();
-    ensureBuilders();
-    SingleFieldBuilderV3<MType, BType, IType> builder =
-        new SingleFieldBuilderV3<MType, BType, IType>(
-            message, this, isClean);
-    messages.add(null);
-    builders.add(builder);
-    onChanged();
-    incrementModCounts();
-    return builder.getBuilder();
-  }
-
-  /**
-   * Inserts a new builder at the specified position in this list.
-   * Shifts the element currently at that position (if any) and any subsequent
-   * elements to the right (adds one to their indices).
-   *
-   * @param index the index at which to insert the builder
-   * @param message the message to add which is the basis of the builder
-   * @return the builder
-   */
-  public BType addBuilder(int index, MType message) {
-    ensureMutableMessageList();
-    ensureBuilders();
-    SingleFieldBuilderV3<MType, BType, IType> builder =
-        new SingleFieldBuilderV3<MType, BType, IType>(
-            message, this, isClean);
-    messages.add(index, null);
-    builders.add(index, builder);
-    onChanged();
-    incrementModCounts();
-    return builder.getBuilder();
-  }
-
-  /**
-   * Removes the element at the specified position in this list. Shifts any
-   * subsequent elements to the left (subtracts one from their indices).
-   * Returns the element that was removed from the list.
-   *
-   * @param index the index at which to remove the message
-   */
-  public void remove(int index) {
-    ensureMutableMessageList();
-    messages.remove(index);
-    if (builders != null) {
-      SingleFieldBuilderV3<MType, BType, IType> entry =
-          builders.remove(index);
-      if (entry != null) {
-        entry.dispose();
-      }
-    }
-    onChanged();
-    incrementModCounts();
-  }
-
-  /**
-   * Removes all of the elements from this list.
-   * The list will be empty after this call returns.
-   */
-  public void clear() {
-    messages = Collections.emptyList();
-    isMessagesListMutable = false;
-    if (builders != null) {
-      for (SingleFieldBuilderV3<MType, BType, IType> entry :
-          builders) {
-        if (entry != null) {
-          entry.dispose();
-        }
-      }
-      builders = null;
-    }
-    onChanged();
-    incrementModCounts();
-  }
-
-  /**
-   * Builds the list of messages from the builder and returns them.
-   *
-   * @return an immutable list of messages
-   */
-  public List<MType> build() {
-    // Now that build has been called, we are required to dispatch
-    // invalidations.
-    isClean = true;
-
-    if (!isMessagesListMutable && builders == null) {
-      // We still have an immutable list and we never created a builder.
-      return messages;
-    }
-
-    boolean allMessagesInSync = true;
-    if (!isMessagesListMutable) {
-      // We still have an immutable list. Let's see if any of them are out
-      // of sync with their builders.
-      for (int i = 0; i < messages.size(); i++) {
-        Message message = messages.get(i);
-        SingleFieldBuilderV3<MType, BType, IType> builder = builders.get(i);
-        if (builder != null) {
-          if (builder.build() != message) {
-            allMessagesInSync = false;
-            break;
-          }
-        }
-      }
-      if (allMessagesInSync) {
-        // Immutable list is still in sync.
-        return messages;
-      }
-    }
-
-    // Need to make sure messages is up to date
-    ensureMutableMessageList();
-    for (int i = 0; i < messages.size(); i++) {
-      messages.set(i, getMessage(i, true));
-    }
-
-    // We're going to return our list as immutable so we mark that we can
-    // no longer update it.
-    messages = Collections.unmodifiableList(messages);
-    isMessagesListMutable = false;
-    return messages;
-  }
-
-  /**
-   * Gets a view of the builder as a list of messages. The returned list is live
-   * and will reflect any changes to the underlying builder.
-   *
-   * @return the messages in the list
-   */
-  public List<MType> getMessageList() {
-    if (externalMessageList == null) {
-      externalMessageList =
-          new MessageExternalList<MType, BType, IType>(this);
-    }
-    return externalMessageList;
-  }
-
-  /**
-   * Gets a view of the builder as a list of builders. This returned list is
-   * live and will reflect any changes to the underlying builder.
-   *
-   * @return the builders in the list
-   */
-  public List<BType> getBuilderList() {
-    if (externalBuilderList == null) {
-      externalBuilderList =
-          new BuilderExternalList<MType, BType, IType>(this);
-    }
-    return externalBuilderList;
-  }
-
-  /**
-   * Gets a view of the builder as a list of MessageOrBuilders. This returned
-   * list is live and will reflect any changes to the underlying builder.
-   *
-   * @return the builders in the list
-   */
-  public List<IType> getMessageOrBuilderList() {
-    if (externalMessageOrBuilderList == null) {
-      externalMessageOrBuilderList =
-          new MessageOrBuilderExternalList<MType, BType, IType>(this);
-    }
-    return externalMessageOrBuilderList;
-  }
-
-  /**
-   * Called when a the builder or one of its nested children has changed
-   * and any parent should be notified of its invalidation.
-   */
-  private void onChanged() {
-    if (isClean && parent != null) {
-      parent.markDirty();
-
-      // Don't keep dispatching invalidations until build is called again.
-      isClean = false;
-    }
-  }
-
-  @Override
-  public void markDirty() {
-    onChanged();
-  }
-
-  /**
-   * Increments the mod counts so that an ConcurrentModificationException can
-   * be thrown if calling code tries to modify the builder while its iterating
-   * the list.
-   */
-  private void incrementModCounts() {
-    if (externalMessageList != null) {
-      externalMessageList.incrementModCount();
-    }
-    if (externalBuilderList != null) {
-      externalBuilderList.incrementModCount();
-    }
-    if (externalMessageOrBuilderList != null) {
-      externalMessageOrBuilderList.incrementModCount();
-    }
-  }
-
-  /**
-   * Provides a live view of the builder as a list of messages.
-   *
-   * @param <MType> the type of message for the field
-   * @param <BType> the type of builder for the field
-   * @param <IType> the common interface for the message and the builder
-   */
-  private static class MessageExternalList<
-      MType extends AbstractMessage,
-      BType extends AbstractMessage.Builder,
-      IType extends MessageOrBuilder>
-      extends AbstractList<MType> implements List<MType> {
-
-    RepeatedFieldBuilderV3<MType, BType, IType> builder;
-
-    MessageExternalList(
-        RepeatedFieldBuilderV3<MType, BType, IType> builder) {
-      this.builder = builder;
-    }
-
-    @Override
-    public int size() {
-      return this.builder.getCount();
-    }
-
-    @Override
-    public MType get(int index) {
-      return builder.getMessage(index);
-    }
-
-    void incrementModCount() {
-      modCount++;
-    }
-  }
-
-  /**
-   * Provides a live view of the builder as a list of builders.
-   *
-   * @param <MType> the type of message for the field
-   * @param <BType> the type of builder for the field
-   * @param <IType> the common interface for the message and the builder
-   */
-  private static class BuilderExternalList<
-      MType extends AbstractMessage,
-      BType extends AbstractMessage.Builder,
-      IType extends MessageOrBuilder>
-      extends AbstractList<BType> implements List<BType> {
-
-    RepeatedFieldBuilderV3<MType, BType, IType> builder;
-
-    BuilderExternalList(
-        RepeatedFieldBuilderV3<MType, BType, IType> builder) {
-      this.builder = builder;
-    }
-
-    @Override
-    public int size() {
-      return this.builder.getCount();
-    }
-
-    @Override
-    public BType get(int index) {
-      return builder.getBuilder(index);
-    }
-
-    void incrementModCount() {
-      modCount++;
-    }
-  }
-
-  /**
-   * Provides a live view of the builder as a list of builders.
-   *
-   * @param <MType> the type of message for the field
-   * @param <BType> the type of builder for the field
-   * @param <IType> the common interface for the message and the builder
-   */
-  private static class MessageOrBuilderExternalList<
-      MType extends AbstractMessage,
-      BType extends AbstractMessage.Builder,
-      IType extends MessageOrBuilder>
-      extends AbstractList<IType> implements List<IType> {
-
-    RepeatedFieldBuilderV3<MType, BType, IType> builder;
-
-    MessageOrBuilderExternalList(
-        RepeatedFieldBuilderV3<MType, BType, IType> builder) {
-      this.builder = builder;
-    }
-
-    @Override
-    public int size() {
-      return this.builder.getCount();
-    }
-
-    @Override
-    public IType get(int index) {
-      return builder.getMessageOrBuilder(index);
-    }
-
-    void incrementModCount() {
-      modCount++;
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RopeByteString.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RopeByteString.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RopeByteString.java
deleted file mode 100644
index a67ed34..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RopeByteString.java
+++ /dev/null
@@ -1,897 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.ByteArrayInputStream;
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.InvalidObjectException;
-import java.io.ObjectInputStream;
-import java.io.OutputStream;
-import java.nio.ByteBuffer;
-import java.nio.charset.Charset;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Iterator;
-import java.util.List;
-import java.util.NoSuchElementException;
-import java.util.Stack;
-
-/**
- * Class to represent {@code ByteStrings} formed by concatenation of other
- * ByteStrings, without copying the data in the pieces. The concatenation is
- * represented as a tree whose leaf nodes are each a
- * {@link org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.LeafByteString}.
- *
- * <p>Most of the operation here is inspired by the now-famous paper <a
- * href="https://web.archive.org/web/20060202015456/http://www.cs.ubc.ca/local/reading/proceedings/spe91-95/spe/vol25/issue12/spe986.pdf">
- * BAP95 </a> Ropes: an Alternative to Strings hans-j. boehm, russ atkinson and
- * michael plass
- *
- * <p>The algorithms described in the paper have been implemented for character
- * strings in {@code com.google.common.string.Rope} and in the c++ class {@code
- * cord.cc}.
- *
- * <p>Fundamentally the Rope algorithm represents the collection of pieces as a
- * binary tree. BAP95 uses a Fibonacci bound relating depth to a minimum
- * sequence length, sequences that are too short relative to their depth cause a
- * tree rebalance.  More precisely, a tree of depth d is "balanced" in the
- * terminology of BAP95 if its length is at least F(d+2), where F(n) is the
- * n-the Fibonacci number. Thus for depths 0, 1, 2, 3, 4, 5,... we have minimum
- * lengths 1, 2, 3, 5, 8, 13,...
- *
- * @author carlanton@google.com (Carl Haverl)
- */
-final class RopeByteString extends ByteString {
-
-  /**
-   * BAP95. Let Fn be the nth Fibonacci number. A {@link RopeByteString} of
-   * depth n is "balanced", i.e flat enough, if its length is at least Fn+2,
-   * e.g. a "balanced" {@link RopeByteString} of depth 1 must have length at
-   * least 2, of depth 4 must have length >= 8, etc.
-   *
-   * <p>There's nothing special about using the Fibonacci numbers for this, but
-   * they are a reasonable sequence for encapsulating the idea that we are OK
-   * with longer strings being encoded in deeper binary trees.
-   *
-   * <p>For 32-bit integers, this array has length 46.
-   */
-  private static final int[] minLengthByDepth;
-
-  static {
-    // Dynamically generate the list of Fibonacci numbers the first time this
-    // class is accessed.
-    List<Integer> numbers = new ArrayList<Integer>();
-
-    // we skip the first Fibonacci number (1).  So instead of: 1 1 2 3 5 8 ...
-    // we have: 1 2 3 5 8 ...
-    int f1 = 1;
-    int f2 = 1;
-
-    // get all the values until we roll over.
-    while (f2 > 0) {
-      numbers.add(f2);
-      int temp = f1 + f2;
-      f1 = f2;
-      f2 = temp;
-    }
-
-    // we include this here so that we can index this array to [x + 1] in the
-    // loops below.
-    numbers.add(Integer.MAX_VALUE);
-    minLengthByDepth = new int[numbers.size()];
-    for (int i = 0; i < minLengthByDepth.length; i++) {
-      // unbox all the values
-      minLengthByDepth[i] = numbers.get(i);
-    }
-  }
-
-  private final int totalLength;
-  private final ByteString left;
-  private final ByteString right;
-  private final int leftLength;
-  private final int treeDepth;
-
-  /**
-   * Create a new RopeByteString, which can be thought of as a new tree node, by
-   * recording references to the two given strings.
-   *
-   * @param left  string on the left of this node, should have {@code size() >
-   *              0}
-   * @param right string on the right of this node, should have {@code size() >
-   *              0}
-   */
-  private RopeByteString(ByteString left, ByteString right) {
-    this.left = left;
-    this.right = right;
-    leftLength = left.size();
-    totalLength = leftLength + right.size();
-    treeDepth = Math.max(left.getTreeDepth(), right.getTreeDepth()) + 1;
-  }
-
-  /**
-   * Concatenate the given strings while performing various optimizations to
-   * slow the growth rate of tree depth and tree node count. The result is
-   * either a {@link org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.LeafByteString} or a
-   * {@link RopeByteString} depending on which optimizations, if any, were
-   * applied.
-   *
-   * <p>Small pieces of length less than {@link
-   * ByteString#CONCATENATE_BY_COPY_SIZE} may be copied by value here, as in
-   * BAP95.  Large pieces are referenced without copy.
-   *
-   * @param left  string on the left
-   * @param right string on the right
-   * @return concatenation representing the same sequence as the given strings
-   */
-  static ByteString concatenate(ByteString left, ByteString right) {
-    if (right.size() == 0) {
-      return left;
-    }
-
-    if (left.size() == 0) {
-      return right;
-    }
-
-    final int newLength = left.size() + right.size();
-    if (newLength < ByteString.CONCATENATE_BY_COPY_SIZE) {
-      // Optimization from BAP95: For short (leaves in paper, but just short
-      // here) total length, do a copy of data to a new leaf.
-      return concatenateBytes(left, right);
-    }
-
-    if (left instanceof RopeByteString) {
-      final RopeByteString leftRope = (RopeByteString) left;
-      if (leftRope.right.size() + right.size() < CONCATENATE_BY_COPY_SIZE) {
-        // Optimization from BAP95: As an optimization of the case where the
-        // ByteString is constructed by repeated concatenate, recognize the case
-        // where a short string is concatenated to a left-hand node whose
-        // right-hand branch is short.  In the paper this applies to leaves, but
-        // we just look at the length here. This has the advantage of shedding
-        // references to unneeded data when substrings have been taken.
-        //
-        // When we recognize this case, we do a copy of the data and create a
-        // new parent node so that the depth of the result is the same as the
-        // given left tree.
-        ByteString newRight = concatenateBytes(leftRope.right, right);
-        return new RopeByteString(leftRope.left, newRight);
-      }
-
-      if (leftRope.left.getTreeDepth() > leftRope.right.getTreeDepth()
-          && leftRope.getTreeDepth() > right.getTreeDepth()) {
-        // Typically for concatenate-built strings the left-side is deeper than
-        // the right.  This is our final attempt to concatenate without
-        // increasing the tree depth.  We'll redo the node on the RHS.  This
-        // is yet another optimization for building the string by repeatedly
-        // concatenating on the right.
-        ByteString newRight = new RopeByteString(leftRope.right, right);
-        return new RopeByteString(leftRope.left, newRight);
-      }
-    }
-
-    // Fine, we'll add a node and increase the tree depth--unless we rebalance ;^)
-    int newDepth = Math.max(left.getTreeDepth(), right.getTreeDepth()) + 1;
-    if (newLength >= minLengthByDepth[newDepth]) {
-      // The tree is shallow enough, so don't rebalance
-      return new RopeByteString(left, right);
-    }
-
-    return new Balancer().balance(left, right);
-  }
-
-  /**
-   * Concatenates two strings by copying data values. This is called in a few
-   * cases in order to reduce the growth of the number of tree nodes.
-   *
-   * @param left  string on the left
-   * @param right string on the right
-   * @return string formed by copying data bytes
-   */
-  private static ByteString concatenateBytes(ByteString left,
-      ByteString right) {
-    int leftSize = left.size();
-    int rightSize = right.size();
-    byte[] bytes = new byte[leftSize + rightSize];
-    left.copyTo(bytes, 0, 0, leftSize);
-    right.copyTo(bytes, 0, leftSize, rightSize);
-    return ByteString.wrap(bytes);  // Constructor wraps bytes
-  }
-
-  /**
-   * Create a new RopeByteString for testing only while bypassing all the
-   * defenses of {@link #concatenate(ByteString, ByteString)}. This allows
-   * testing trees of specific structure. We are also able to insert empty
-   * leaves, though these are dis-allowed, so that we can make sure the
-   * implementation can withstand their presence.
-   *
-   * @param left  string on the left of this node
-   * @param right string on the right of this node
-   * @return an unsafe instance for testing only
-   */
-  static RopeByteString newInstanceForTest(ByteString left, ByteString right) {
-    return new RopeByteString(left, right);
-  }
-
-  /**
-   * Gets the byte at the given index.
-   * Throws {@link ArrayIndexOutOfBoundsException} for backwards-compatibility
-   * reasons although it would more properly be {@link
-   * IndexOutOfBoundsException}.
-   *
-   * @param index index of byte
-   * @return the value
-   * @throws ArrayIndexOutOfBoundsException {@code index} is < 0 or >= size
-   */
-  @Override
-  public byte byteAt(int index) {
-    checkIndex(index, totalLength);
-
-    // Find the relevant piece by recursive descent
-    if (index < leftLength) {
-      return left.byteAt(index);
-    }
-
-    return right.byteAt(index - leftLength);
-  }
-
-  @Override
-  public int size() {
-    return totalLength;
-  }
-
-  // =================================================================
-  // Pieces
-
-  @Override
-  protected int getTreeDepth() {
-    return treeDepth;
-  }
-
-  /**
-   * Determines if the tree is balanced according to BAP95, which means the tree
-   * is flat-enough with respect to the bounds. Note that this definition of
-   * balanced is one where sub-trees of balanced trees are not necessarily
-   * balanced.
-   *
-   * @return true if the tree is balanced
-   */
-  @Override
-  protected boolean isBalanced() {
-    return totalLength >= minLengthByDepth[treeDepth];
-  }
-
-  /**
-   * Takes a substring of this one. This involves recursive descent along the
-   * left and right edges of the substring, and referencing any wholly contained
-   * segments in between. Any leaf nodes entirely uninvolved in the substring
-   * will not be referenced by the substring.
-   *
-   * <p>Substrings of {@code length < 2} should result in at most a single
-   * recursive call chain, terminating at a leaf node. Thus the result will be a
-   * {@link org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.LeafByteString}.
-   *
-   * @param beginIndex start at this index
-   * @param endIndex   the last character is the one before this index
-   * @return substring leaf node or tree
-   */
-  @Override
-  public ByteString substring(int beginIndex, int endIndex) {
-    final int length = checkRange(beginIndex, endIndex, totalLength);
-
-    if (length == 0) {
-      // Empty substring
-      return ByteString.EMPTY;
-    }
-
-    if (length == totalLength) {
-      // The whole string
-      return this;
-    }
-
-    // Proper substring
-    if (endIndex <= leftLength) {
-      // Substring on the left
-      return left.substring(beginIndex, endIndex);
-    }
-
-    if (beginIndex >= leftLength) {
-      // Substring on the right
-      return right.substring(beginIndex - leftLength, endIndex - leftLength);
-    }
-
-    // Split substring
-    ByteString leftSub = left.substring(beginIndex);
-    ByteString rightSub = right.substring(0, endIndex - leftLength);
-    // Intentionally not rebalancing, since in many cases these two
-    // substrings will already be less deep than the top-level
-    // RopeByteString we're taking a substring of.
-    return new RopeByteString(leftSub, rightSub);
-  }
-
-  // =================================================================
-  // ByteString -> byte[]
-
-  @Override
-  protected void copyToInternal(byte[] target, int sourceOffset,
-      int targetOffset, int numberToCopy) {
-   if (sourceOffset + numberToCopy <= leftLength) {
-      left.copyToInternal(target, sourceOffset, targetOffset, numberToCopy);
-    } else if (sourceOffset >= leftLength) {
-      right.copyToInternal(target, sourceOffset - leftLength, targetOffset,
-          numberToCopy);
-    } else {
-      int leftLength = this.leftLength - sourceOffset;
-      left.copyToInternal(target, sourceOffset, targetOffset, leftLength);
-      right.copyToInternal(target, 0, targetOffset + leftLength,
-          numberToCopy - leftLength);
-    }
-  }
-
-  @Override
-  public void copyTo(ByteBuffer target) {
-    left.copyTo(target);
-    right.copyTo(target);
-  }
-
-  @Override
-  public ByteBuffer asReadOnlyByteBuffer() {
-    ByteBuffer byteBuffer = ByteBuffer.wrap(toByteArray());
-    return byteBuffer.asReadOnlyBuffer();
-  }
-
-  @Override
-  public List<ByteBuffer> asReadOnlyByteBufferList() {
-    // Walk through the list of LeafByteString's that make up this
-    // rope, and add each one as a read-only ByteBuffer.
-    List<ByteBuffer> result = new ArrayList<ByteBuffer>();
-    PieceIterator pieces = new PieceIterator(this);
-    while (pieces.hasNext()) {
-      LeafByteString byteString = pieces.next();
-      result.add(byteString.asReadOnlyByteBuffer());
-    }
-    return result;
-  }
-
-  @Override
-  public void writeTo(OutputStream outputStream) throws IOException {
-    left.writeTo(outputStream);
-    right.writeTo(outputStream);
-  }
-
-  @Override
-  void writeToInternal(OutputStream out, int sourceOffset,
-      int numberToWrite) throws IOException {
-    if (sourceOffset + numberToWrite <= leftLength) {
-      left.writeToInternal(out, sourceOffset, numberToWrite);
-    } else if (sourceOffset >= leftLength) {
-      right.writeToInternal(out, sourceOffset - leftLength, numberToWrite);
-    } else {
-      int numberToWriteInLeft = leftLength - sourceOffset;
-      left.writeToInternal(out, sourceOffset, numberToWriteInLeft);
-      right.writeToInternal(out, 0, numberToWrite - numberToWriteInLeft);
-    }
-  }
-
-  @Override
-  void writeTo(ByteOutput output) throws IOException {
-    left.writeTo(output);
-    right.writeTo(output);
-  }
-
-
-  @Override
-  protected String toStringInternal(Charset charset) {
-    return new String(toByteArray(), charset);
-  }
-
-  // =================================================================
-  // UTF-8 decoding
-
-  @Override
-  public boolean isValidUtf8() {
-    int leftPartial = left.partialIsValidUtf8(Utf8.COMPLETE, 0, leftLength);
-    int state = right.partialIsValidUtf8(leftPartial, 0, right.size());
-    return state == Utf8.COMPLETE;
-  }
-
-  @Override
-  protected int partialIsValidUtf8(int state, int offset, int length) {
-    int toIndex = offset + length;
-    if (toIndex <= leftLength) {
-      return left.partialIsValidUtf8(state, offset, length);
-    } else if (offset >= leftLength) {
-      return right.partialIsValidUtf8(state, offset - leftLength, length);
-    } else {
-      int leftLength = this.leftLength - offset;
-      int leftPartial = left.partialIsValidUtf8(state, offset, leftLength);
-      return right.partialIsValidUtf8(leftPartial, 0, length - leftLength);
-    }
-  }
-
-  // =================================================================
-  // equals() and hashCode()
-
-  @Override
-  public boolean equals(Object other) {
-    if (other == this) {
-      return true;
-    }
-    if (!(other instanceof ByteString)) {
-      return false;
-    }
-
-    ByteString otherByteString = (ByteString) other;
-    if (totalLength != otherByteString.size()) {
-      return false;
-    }
-    if (totalLength == 0) {
-      return true;
-    }
-
-    // You don't really want to be calling equals on long strings, but since
-    // we cache the hashCode, we effectively cache inequality. We use the cached
-    // hashCode if it's already computed.  It's arguable we should compute the
-    // hashCode here, and if we're going to be testing a bunch of byteStrings,
-    // it might even make sense.
-    int thisHash = peekCachedHashCode();
-    int thatHash = otherByteString.peekCachedHashCode();
-    if (thisHash != 0 && thatHash != 0 && thisHash != thatHash) {
-      return false;
-    }
-
-    return equalsFragments(otherByteString);
-  }
-
-  /**
-   * Determines if this string is equal to another of the same length by
-   * iterating over the leaf nodes. On each step of the iteration, the
-   * overlapping segments of the leaves are compared.
-   *
-   * @param other string of the same length as this one
-   * @return true if the values of this string equals the value of the given
-   *         one
-   */
-  private boolean equalsFragments(ByteString other) {
-    int thisOffset = 0;
-    Iterator<LeafByteString> thisIter = new PieceIterator(this);
-    LeafByteString thisString = thisIter.next();
-
-    int thatOffset = 0;
-    Iterator<LeafByteString> thatIter = new PieceIterator(other);
-    LeafByteString thatString = thatIter.next();
-
-    int pos = 0;
-    while (true) {
-      int thisRemaining = thisString.size() - thisOffset;
-      int thatRemaining = thatString.size() - thatOffset;
-      int bytesToCompare = Math.min(thisRemaining, thatRemaining);
-
-      // At least one of the offsets will be zero
-      boolean stillEqual = (thisOffset == 0)
-          ? thisString.equalsRange(thatString, thatOffset, bytesToCompare)
-          : thatString.equalsRange(thisString, thisOffset, bytesToCompare);
-      if (!stillEqual) {
-        return false;
-      }
-
-      pos += bytesToCompare;
-      if (pos >= totalLength) {
-        if (pos == totalLength) {
-          return true;
-        }
-        throw new IllegalStateException();
-      }
-      // We always get to the end of at least one of the pieces
-      if (bytesToCompare == thisRemaining) { // If reached end of this
-        thisOffset = 0;
-        thisString = thisIter.next();
-      } else {
-        thisOffset += bytesToCompare;
-      }
-      if (bytesToCompare == thatRemaining) { // If reached end of that
-        thatOffset = 0;
-        thatString = thatIter.next();
-      } else {
-        thatOffset += bytesToCompare;
-      }
-    }
-  }
-
-  @Override
-  protected int partialHash(int h, int offset, int length) {
-    int toIndex = offset + length;
-    if (toIndex <= leftLength) {
-      return left.partialHash(h, offset, length);
-    } else if (offset >= leftLength) {
-      return right.partialHash(h, offset - leftLength, length);
-    } else {
-      int leftLength = this.leftLength - offset;
-      int leftPartial = left.partialHash(h, offset, leftLength);
-      return right.partialHash(leftPartial, 0, length - leftLength);
-    }
-  }
-
-  // =================================================================
-  // Input stream
-
-  @Override
-  public CodedInputStream newCodedInput() {
-    return CodedInputStream.newInstance(new RopeInputStream());
-  }
-
-  @Override
-  public InputStream newInput() {
-    return new RopeInputStream();
-  }
-
-  /**
-   * This class implements the balancing algorithm of BAP95. In the paper the
-   * authors use an array to keep track of pieces, while here we use a stack.
-   * The tree is balanced by traversing subtrees in left to right order, and the
-   * stack always contains the part of the string we've traversed so far.
-   *
-   * <p>One surprising aspect of the algorithm is the result of balancing is not
-   * necessarily balanced, though it is nearly balanced.  For details, see
-   * BAP95.
-   */
-  private static class Balancer {
-    // Stack containing the part of the string, starting from the left, that
-    // we've already traversed.  The final string should be the equivalent of
-    // concatenating the strings on the stack from bottom to top.
-    private final Stack<ByteString> prefixesStack = new Stack<ByteString>();
-
-    private ByteString balance(ByteString left, ByteString right) {
-      doBalance(left);
-      doBalance(right);
-
-      // Sweep stack to gather the result
-      ByteString partialString = prefixesStack.pop();
-      while (!prefixesStack.isEmpty()) {
-        ByteString newLeft = prefixesStack.pop();
-        partialString = new RopeByteString(newLeft, partialString);
-      }
-      // We should end up with a RopeByteString since at a minimum we will
-      // create one from concatenating left and right
-      return partialString;
-    }
-
-    private void doBalance(ByteString root) {
-      // BAP95: Insert balanced subtrees whole. This means the result might not
-      // be balanced, leading to repeated rebalancings on concatenate. However,
-      // these rebalancings are shallow due to ignoring balanced subtrees, and
-      // relatively few calls to insert() result.
-      if (root.isBalanced()) {
-        insert(root);
-      } else if (root instanceof RopeByteString) {
-        RopeByteString rbs = (RopeByteString) root;
-        doBalance(rbs.left);
-        doBalance(rbs.right);
-      } else {
-        throw new IllegalArgumentException(
-            "Has a new type of ByteString been created? Found " +
-                root.getClass());
-      }
-    }
-
-    /**
-     * Push a string on the balance stack (BAP95).  BAP95 uses an array and
-     * calls the elements in the array 'bins'.  We instead use a stack, so the
-     * 'bins' of lengths are represented by differences between the elements of
-     * minLengthByDepth.
-     *
-     * <p>If the length bin for our string, and all shorter length bins, are
-     * empty, we just push it on the stack.  Otherwise, we need to start
-     * concatenating, putting the given string in the "middle" and continuing
-     * until we land in an empty length bin that matches the length of our
-     * concatenation.
-     *
-     * @param byteString string to place on the balance stack
-     */
-    private void insert(ByteString byteString) {
-      int depthBin = getDepthBinForLength(byteString.size());
-      int binEnd = minLengthByDepth[depthBin + 1];
-
-      // BAP95: Concatenate all trees occupying bins representing the length of
-      // our new piece or of shorter pieces, to the extent that is possible.
-      // The goal is to clear the bin which our piece belongs in, but that may
-      // not be entirely possible if there aren't enough longer bins occupied.
-      if (prefixesStack.isEmpty() || prefixesStack.peek().size() >= binEnd) {
-        prefixesStack.push(byteString);
-      } else {
-        int binStart = minLengthByDepth[depthBin];
-
-        // Concatenate the subtrees of shorter length
-        ByteString newTree = prefixesStack.pop();
-        while (!prefixesStack.isEmpty()
-            && prefixesStack.peek().size() < binStart) {
-          ByteString left = prefixesStack.pop();
-          newTree = new RopeByteString(left, newTree);
-        }
-
-        // Concatenate the given string
-        newTree = new RopeByteString(newTree, byteString);
-
-        // Continue concatenating until we land in an empty bin
-        while (!prefixesStack.isEmpty()) {
-          depthBin = getDepthBinForLength(newTree.size());
-          binEnd = minLengthByDepth[depthBin + 1];
-          if (prefixesStack.peek().size() < binEnd) {
-            ByteString left = prefixesStack.pop();
-            newTree = new RopeByteString(left, newTree);
-          } else {
-            break;
-          }
-        }
-        prefixesStack.push(newTree);
-      }
-    }
-
-    private int getDepthBinForLength(int length) {
-      int depth = Arrays.binarySearch(minLengthByDepth, length);
-      if (depth < 0) {
-        // It wasn't an exact match, so convert to the index of the containing
-        // fragment, which is one less even than the insertion point.
-        int insertionPoint = -(depth + 1);
-        depth = insertionPoint - 1;
-      }
-
-      return depth;
-    }
-  }
-
-  /**
-   * This class is a continuable tree traversal, which keeps the state
-   * information which would exist on the stack in a recursive traversal instead
-   * on a stack of "Bread Crumbs". The maximum depth of the stack in this
-   * iterator is the same as the depth of the tree being traversed.
-   *
-   * <p>This iterator is used to implement
-   * {@link RopeByteString#equalsFragments(ByteString)}.
-   */
-  private static class PieceIterator implements Iterator<LeafByteString> {
-
-    private final Stack<RopeByteString> breadCrumbs =
-        new Stack<RopeByteString>();
-    private LeafByteString next;
-
-    private PieceIterator(ByteString root) {
-      next = getLeafByLeft(root);
-    }
-
-    private LeafByteString getLeafByLeft(ByteString root) {
-      ByteString pos = root;
-      while (pos instanceof RopeByteString) {
-        RopeByteString rbs = (RopeByteString) pos;
-        breadCrumbs.push(rbs);
-        pos = rbs.left;
-      }
-      return (LeafByteString) pos;
-    }
-
-    private LeafByteString getNextNonEmptyLeaf() {
-      while (true) {
-        // Almost always, we go through this loop exactly once.  However, if
-        // we discover an empty string in the rope, we toss it and try again.
-        if (breadCrumbs.isEmpty()) {
-          return null;
-        } else {
-          LeafByteString result = getLeafByLeft(breadCrumbs.pop().right);
-          if (!result.isEmpty()) {
-            return result;
-          }
-        }
-      }
-    }
-
-    @Override
-    public boolean hasNext() {
-      return next != null;
-    }
-
-    /**
-     * Returns the next item and advances one
-     * {@link org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.LeafByteString}.
-     *
-     * @return next non-empty LeafByteString or {@code null}
-     */
-    @Override
-    public LeafByteString next() {
-      if (next == null) {
-        throw new NoSuchElementException();
-      }
-      LeafByteString result = next;
-      next = getNextNonEmptyLeaf();
-      return result;
-    }
-
-    @Override
-    public void remove() {
-      throw new UnsupportedOperationException();
-    }
-  }
-
-  // =================================================================
-  // Serializable
-
-  private static final long serialVersionUID = 1L;
-
-  Object writeReplace() {
-    return ByteString.wrap(toByteArray());
-  }
-
-  private void readObject(@SuppressWarnings("unused") ObjectInputStream in) throws IOException {
-    throw new InvalidObjectException(
-        "RopeByteStream instances are not to be serialized directly");
-  }
-
-  /**
-   * This class is the {@link RopeByteString} equivalent for
-   * {@link ByteArrayInputStream}.
-   */
-  private class RopeInputStream extends InputStream {
-    // Iterates through the pieces of the rope
-    private PieceIterator pieceIterator;
-    // The current piece
-    private LeafByteString currentPiece;
-    // The size of the current piece
-    private int currentPieceSize;
-    // The index of the next byte to read in the current piece
-    private int currentPieceIndex;
-    // The offset of the start of the current piece in the rope byte string
-    private int currentPieceOffsetInRope;
-    // Offset in the buffer at which user called mark();
-    private int mark;
-
-    public RopeInputStream() {
-      initialize();
-    }
-
-    @Override
-    public int read(byte b[], int offset, int length)  {
-      if (b == null) {
-        throw new NullPointerException();
-      } else if (offset < 0 || length < 0 || length > b.length - offset) {
-        throw new IndexOutOfBoundsException();
-      }
-      return readSkipInternal(b, offset, length);
-    }
-
-    @Override
-    public long skip(long length) {
-      if (length < 0) {
-        throw new IndexOutOfBoundsException();
-      } else if (length > Integer.MAX_VALUE) {
-        length = Integer.MAX_VALUE;
-      }
-      return readSkipInternal(null, 0, (int) length);
-    }
-
-    /**
-     * Internal implementation of read and skip.  If b != null, then read the
-     * next {@code length} bytes into the buffer {@code b} at
-     * offset {@code offset}.  If b == null, then skip the next {@code length}
-     * bytes.
-     * <p>
-     * This method assumes that all error checking has already happened.
-     * <p>
-     * Returns the actual number of bytes read or skipped.
-     */
-    private int readSkipInternal(byte b[], int offset, int length)  {
-      int bytesRemaining = length;
-      while (bytesRemaining > 0) {
-        advanceIfCurrentPieceFullyRead();
-        if (currentPiece == null) {
-          if (bytesRemaining == length) {
-             // We didn't manage to read anything
-             return -1;
-           }
-          break;
-        } else {
-          // Copy the bytes from this piece.
-          int currentPieceRemaining = currentPieceSize - currentPieceIndex;
-          int count = Math.min(currentPieceRemaining, bytesRemaining);
-          if (b != null) {
-            currentPiece.copyTo(b, currentPieceIndex, offset, count);
-            offset += count;
-          }
-          currentPieceIndex += count;
-          bytesRemaining -= count;
-        }
-      }
-       // Return the number of bytes read.
-      return length - bytesRemaining;
-    }
-
-    @Override
-    public int read() throws IOException {
-      advanceIfCurrentPieceFullyRead();
-      if (currentPiece == null) {
-        return -1;
-      } else {
-        return currentPiece.byteAt(currentPieceIndex++) & 0xFF;
-      }
-    }
-
-    @Override
-    public int available() throws IOException {
-      int bytesRead = currentPieceOffsetInRope + currentPieceIndex;
-      return RopeByteString.this.size() - bytesRead;
-    }
-
-    @Override
-    public boolean markSupported() {
-      return true;
-    }
-
-    @Override
-    public void mark(int readAheadLimit) {
-      // Set the mark to our position in the byte string
-      mark = currentPieceOffsetInRope + currentPieceIndex;
-    }
-
-    @Override
-    public synchronized void reset() {
-      // Just reinitialize and skip the specified number of bytes.
-      initialize();
-      readSkipInternal(null, 0, mark);
-    }
-
-    /** Common initialization code used by both the constructor and reset() */
-    private void initialize() {
-      pieceIterator = new PieceIterator(RopeByteString.this);
-      currentPiece = pieceIterator.next();
-      currentPieceSize = currentPiece.size();
-      currentPieceIndex = 0;
-      currentPieceOffsetInRope = 0;
-    }
-
-    /**
-     * Skips to the next piece if we have read all the data in the current
-     * piece.  Sets currentPiece to null if we have reached the end of the
-     * input.
-     */
-    private void advanceIfCurrentPieceFullyRead() {
-      if (currentPiece != null && currentPieceIndex == currentPieceSize) {
-        // Generally, we can only go through this loop at most once, since
-        // empty strings can't end up in a rope.  But better to test.
-        currentPieceOffsetInRope += currentPieceSize;
-        currentPieceIndex = 0;
-        if (pieceIterator.hasNext()) {
-          currentPiece = pieceIterator.next();
-          currentPieceSize = currentPiece.size();
-        } else {
-          currentPiece = null;
-          currentPieceSize = 0;
-        }
-      }
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcCallback.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcCallback.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcCallback.java
deleted file mode 100644
index dc30060..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcCallback.java
+++ /dev/null
@@ -1,47 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Interface for an RPC callback, normally called when an RPC completes.
- * {@code ParameterType} is normally the method's response message type.
- *
- * <p>Starting with version 2.3.0, RPC implementations should not try to build
- * on this, but should instead provide code generator plugins which generate
- * code specific to the particular RPC implementation.  This way the generated
- * code can be more appropriate for the implementation in use and can avoid
- * unnecessary layers of indirection.
- *
- * @author kenton@google.com Kenton Varda
- */
-public interface RpcCallback<ParameterType> {
-  void run(ParameterType parameter);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcChannel.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcChannel.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcChannel.java
deleted file mode 100644
index bf10784..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcChannel.java
+++ /dev/null
@@ -1,71 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <p>Abstract interface for an RPC channel.  An {@code RpcChannel} represents a
- * communication line to a {@link Service} which can be used to call that
- * {@link Service}'s methods.  The {@link Service} may be running on another
- * machine.  Normally, you should not call an {@code RpcChannel} directly, but
- * instead construct a stub {@link Service} wrapping it.  Example:
- *
- * <pre>
- *   RpcChannel channel = rpcImpl.newChannel("remotehost.example.com:1234");
- *   RpcController controller = rpcImpl.newController();
- *   MyService service = MyService.newStub(channel);
- *   service.myMethod(controller, request, callback);
- * </pre>
- *
- * <p>Starting with version 2.3.0, RPC implementations should not try to build
- * on this, but should instead provide code generator plugins which generate
- * code specific to the particular RPC implementation.  This way the generated
- * code can be more appropriate for the implementation in use and can avoid
- * unnecessary layers of indirection.
- *
- * @author kenton@google.com Kenton Varda
- */
-public interface RpcChannel {
-  /**
-   * Call the given method of the remote service.  This method is similar to
-   * {@code Service.callMethod()} with one important difference:  the caller
-   * decides the types of the {@code Message} objects, not the callee.  The
-   * request may be of any type as long as
-   * {@code request.getDescriptor() == method.getInputType()}.
-   * The response passed to the callback will be of the same type as
-   * {@code responsePrototype} (which must have
-   * {@code getDescriptor() == method.getOutputType()}).
-   */
-  void callMethod(Descriptors.MethodDescriptor method,
-                  RpcController controller,
-                  Message request,
-                  Message responsePrototype,
-                  RpcCallback<Message> done);
-}


[05/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java
deleted file mode 100644
index a30032d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProtos.java
+++ /dev/null
@@ -1,2987 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: test.proto
-
-package org.apache.hadoop.hbase.shaded.ipc.protobuf.generated;
-
-public final class TestProtos {
-  private TestProtos() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  public interface EmptyRequestProtoOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:EmptyRequestProto)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-  }
-  /**
-   * Protobuf type {@code EmptyRequestProto}
-   */
-  public  static final class EmptyRequestProto extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:EmptyRequestProto)
-      EmptyRequestProtoOrBuilder {
-    // Use EmptyRequestProto.newBuilder() to construct.
-    private EmptyRequestProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private EmptyRequestProto() {
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private EmptyRequestProto(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.Builder.class);
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto other = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto) obj;
-
-      boolean result = true;
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code EmptyRequestProto}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:EmptyRequestProto)
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProtoOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyRequestProto_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto build() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto buildPartial() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto result = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto(this);
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto other) {
-        if (other == org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance()) return this;
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:EmptyRequestProto)
-    }
-
-    // @@protoc_insertion_point(class_scope:EmptyRequestProto)
-    private static final org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyRequestProto>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EmptyRequestProto>() {
-      public EmptyRequestProto parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new EmptyRequestProto(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyRequestProto> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyRequestProto> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface EmptyResponseProtoOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:EmptyResponseProto)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-  }
-  /**
-   * Protobuf type {@code EmptyResponseProto}
-   */
-  public  static final class EmptyResponseProto extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:EmptyResponseProto)
-      EmptyResponseProtoOrBuilder {
-    // Use EmptyResponseProto.newBuilder() to construct.
-    private EmptyResponseProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private EmptyResponseProto() {
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private EmptyResponseProto(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.Builder.class);
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto other = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) obj;
-
-      boolean result = true;
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code EmptyResponseProto}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:EmptyResponseProto)
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProtoOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EmptyResponseProto_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto build() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto buildPartial() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto result = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto(this);
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto other) {
-        if (other == org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()) return this;
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:EmptyResponseProto)
-    }
-
-    // @@protoc_insertion_point(class_scope:EmptyResponseProto)
-    private static final org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyResponseProto>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EmptyResponseProto>() {
-      public EmptyResponseProto parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new EmptyResponseProto(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyResponseProto> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EmptyResponseProto> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface EchoRequestProtoOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:EchoRequestProto)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>required string message = 1;</code>
-     */
-    boolean hasMessage();
-    /**
-     * <code>required string message = 1;</code>
-     */
-    java.lang.String getMessage();
-    /**
-     * <code>required string message = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getMessageBytes();
-  }
-  /**
-   * Protobuf type {@code EchoRequestProto}
-   */
-  public  static final class EchoRequestProto extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:EchoRequestProto)
-      EchoRequestProtoOrBuilder {
-    // Use EchoRequestProto.newBuilder() to construct.
-    private EchoRequestProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private EchoRequestProto() {
-      message_ = "";
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private EchoRequestProto(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000001;
-              message_ = bs;
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int MESSAGE_FIELD_NUMBER = 1;
-    private volatile java.lang.Object message_;
-    /**
-     * <code>required string message = 1;</code>
-     */
-    public boolean hasMessage() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required string message = 1;</code>
-     */
-    public java.lang.String getMessage() {
-      java.lang.Object ref = message_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          message_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <code>required string message = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getMessageBytes() {
-      java.lang.Object ref = message_;
-      if (ref instanceof java.lang.String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        message_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (!hasMessage()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, message_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, message_);
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto other = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto) obj;
-
-      boolean result = true;
-      result = result && (hasMessage() == other.hasMessage());
-      if (hasMessage()) {
-        result = result && getMessage()
-            .equals(other.getMessage());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasMessage()) {
-        hash = (37 * hash) + MESSAGE_FIELD_NUMBER;
-        hash = (53 * hash) + getMessage().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code EchoRequestProto}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:EchoRequestProto)
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProtoOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        message_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoRequestProto_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto build() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto buildPartial() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto result = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.message_ = message_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto other) {
-        if (other == org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.getDefaultInstance()) return this;
-        if (other.hasMessage()) {
-          bitField0_ |= 0x00000001;
-          message_ = other.message_;
-          onChanged();
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasMessage()) {
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private java.lang.Object message_ = "";
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public boolean hasMessage() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public java.lang.String getMessage() {
-        java.lang.Object ref = message_;
-        if (!(ref instanceof java.lang.String)) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-              (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-          java.lang.String s = bs.toStringUtf8();
-          if (bs.isValidUtf8()) {
-            message_ = s;
-          }
-          return s;
-        } else {
-          return (java.lang.String) ref;
-        }
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-          getMessageBytes() {
-        java.lang.Object ref = message_;
-        if (ref instanceof String) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                  (java.lang.String) ref);
-          message_ = b;
-          return b;
-        } else {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        }
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public Builder setMessage(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        message_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public Builder clearMessage() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        message_ = getDefaultInstance().getMessage();
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public Builder setMessageBytes(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        message_ = value;
-        onChanged();
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:EchoRequestProto)
-    }
-
-    // @@protoc_insertion_point(class_scope:EchoRequestProto)
-    private static final org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoRequestProto>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EchoRequestProto>() {
-      public EchoRequestProto parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new EchoRequestProto(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoRequestProto> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoRequestProto> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface EchoResponseProtoOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:EchoResponseProto)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>required string message = 1;</code>
-     */
-    boolean hasMessage();
-    /**
-     * <code>required string message = 1;</code>
-     */
-    java.lang.String getMessage();
-    /**
-     * <code>required string message = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getMessageBytes();
-  }
-  /**
-   * Protobuf type {@code EchoResponseProto}
-   */
-  public  static final class EchoResponseProto extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:EchoResponseProto)
-      EchoResponseProtoOrBuilder {
-    // Use EchoResponseProto.newBuilder() to construct.
-    private EchoResponseProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private EchoResponseProto() {
-      message_ = "";
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private EchoResponseProto(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000001;
-              message_ = bs;
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int MESSAGE_FIELD_NUMBER = 1;
-    private volatile java.lang.Object message_;
-    /**
-     * <code>required string message = 1;</code>
-     */
-    public boolean hasMessage() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required string message = 1;</code>
-     */
-    public java.lang.String getMessage() {
-      java.lang.Object ref = message_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          message_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <code>required string message = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getMessageBytes() {
-      java.lang.Object ref = message_;
-      if (ref instanceof java.lang.String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        message_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (!hasMessage()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, message_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, message_);
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto other = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto) obj;
-
-      boolean result = true;
-      result = result && (hasMessage() == other.hasMessage());
-      if (hasMessage()) {
-        result = result && getMessage()
-            .equals(other.getMessage());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasMessage()) {
-        hash = (37 * hash) + MESSAGE_FIELD_NUMBER;
-        hash = (53 * hash) + getMessage().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code EchoResponseProto}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:EchoResponseProto)
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProtoOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        message_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_EchoResponseProto_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto build() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto buildPartial() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto result = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.message_ = message_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto other) {
-        if (other == org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance()) return this;
-        if (other.hasMessage()) {
-          bitField0_ |= 0x00000001;
-          message_ = other.message_;
-          onChanged();
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasMessage()) {
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private java.lang.Object message_ = "";
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public boolean hasMessage() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public java.lang.String getMessage() {
-        java.lang.Object ref = message_;
-        if (!(ref instanceof java.lang.String)) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-              (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-          java.lang.String s = bs.toStringUtf8();
-          if (bs.isValidUtf8()) {
-            message_ = s;
-          }
-          return s;
-        } else {
-          return (java.lang.String) ref;
-        }
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-          getMessageBytes() {
-        java.lang.Object ref = message_;
-        if (ref instanceof String) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                  (java.lang.String) ref);
-          message_ = b;
-          return b;
-        } else {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        }
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public Builder setMessage(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        message_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public Builder clearMessage() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        message_ = getDefaultInstance().getMessage();
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string message = 1;</code>
-       */
-      public Builder setMessageBytes(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        message_ = value;
-        onChanged();
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:EchoResponseProto)
-    }
-
-    // @@protoc_insertion_point(class_scope:EchoResponseProto)
-    private static final org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoResponseProto>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EchoResponseProto>() {
-      public EchoResponseProto parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new EchoResponseProto(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoResponseProto> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EchoResponseProto> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface PauseRequestProtoOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:PauseRequestProto)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>required uint32 ms = 1;</code>
-     */
-    boolean hasMs();
-    /**
-     * <code>required uint32 ms = 1;</code>
-     */
-    int getMs();
-  }
-  /**
-   * Protobuf type {@code PauseRequestProto}
-   */
-  public  static final class PauseRequestProto extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:PauseRequestProto)
-      PauseRequestProtoOrBuilder {
-    // Use PauseRequestProto.newBuilder() to construct.
-    private PauseRequestProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private PauseRequestProto() {
-      ms_ = 0;
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private PauseRequestProto(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              ms_ = input.readUInt32();
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_PauseRequestProto_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.internal_static_PauseRequestProto_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int MS_FIELD_NUMBER = 1;
-    private int ms_;
-    /**
-     * <code>required uint32 ms = 1;</code>
-     */
-    public boolean hasMs() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required uint32 ms = 1;</code>
-     */
-    public int getMs() {
-      return ms_;
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (!hasMs()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeUInt32(1, ms_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.c

<TRUNCATED>

[29/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Field.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Field.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Field.java
deleted file mode 100644
index d33fd75..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Field.java
+++ /dev/null
@@ -1,2450 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * A single field of a message type.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Field}
- */
-public  final class Field extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Field)
-    FieldOrBuilder {
-  // Use Field.newBuilder() to construct.
-  private Field(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Field() {
-    kind_ = 0;
-    cardinality_ = 0;
-    number_ = 0;
-    name_ = "";
-    typeUrl_ = "";
-    oneofIndex_ = 0;
-    packed_ = false;
-    options_ = java.util.Collections.emptyList();
-    jsonName_ = "";
-    defaultValue_ = "";
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Field(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-            int rawValue = input.readEnum();
-
-            kind_ = rawValue;
-            break;
-          }
-          case 16: {
-            int rawValue = input.readEnum();
-
-            cardinality_ = rawValue;
-            break;
-          }
-          case 24: {
-
-            number_ = input.readInt32();
-            break;
-          }
-          case 34: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            name_ = s;
-            break;
-          }
-          case 50: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            typeUrl_ = s;
-            break;
-          }
-          case 56: {
-
-            oneofIndex_ = input.readInt32();
-            break;
-          }
-          case 64: {
-
-            packed_ = input.readBool();
-            break;
-          }
-          case 74: {
-            if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) {
-              options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>();
-              mutable_bitField0_ |= 0x00000080;
-            }
-            options_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.parser(), extensionRegistry));
-            break;
-          }
-          case 82: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            jsonName_ = s;
-            break;
-          }
-          case 90: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            defaultValue_ = s;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) {
-        options_ = java.util.Collections.unmodifiableList(options_);
-      }
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Field_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Field_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder.class);
-  }
-
-  /**
-   * <pre>
-   * Basic field types.
-   * </pre>
-   *
-   * Protobuf enum {@code google.protobuf.Field.Kind}
-   */
-  public enum Kind
-      implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
-    /**
-     * <pre>
-     * Field type unknown.
-     * </pre>
-     *
-     * <code>TYPE_UNKNOWN = 0;</code>
-     */
-    TYPE_UNKNOWN(0),
-    /**
-     * <pre>
-     * Field type double.
-     * </pre>
-     *
-     * <code>TYPE_DOUBLE = 1;</code>
-     */
-    TYPE_DOUBLE(1),
-    /**
-     * <pre>
-     * Field type float.
-     * </pre>
-     *
-     * <code>TYPE_FLOAT = 2;</code>
-     */
-    TYPE_FLOAT(2),
-    /**
-     * <pre>
-     * Field type int64.
-     * </pre>
-     *
-     * <code>TYPE_INT64 = 3;</code>
-     */
-    TYPE_INT64(3),
-    /**
-     * <pre>
-     * Field type uint64.
-     * </pre>
-     *
-     * <code>TYPE_UINT64 = 4;</code>
-     */
-    TYPE_UINT64(4),
-    /**
-     * <pre>
-     * Field type int32.
-     * </pre>
-     *
-     * <code>TYPE_INT32 = 5;</code>
-     */
-    TYPE_INT32(5),
-    /**
-     * <pre>
-     * Field type fixed64.
-     * </pre>
-     *
-     * <code>TYPE_FIXED64 = 6;</code>
-     */
-    TYPE_FIXED64(6),
-    /**
-     * <pre>
-     * Field type fixed32.
-     * </pre>
-     *
-     * <code>TYPE_FIXED32 = 7;</code>
-     */
-    TYPE_FIXED32(7),
-    /**
-     * <pre>
-     * Field type bool.
-     * </pre>
-     *
-     * <code>TYPE_BOOL = 8;</code>
-     */
-    TYPE_BOOL(8),
-    /**
-     * <pre>
-     * Field type string.
-     * </pre>
-     *
-     * <code>TYPE_STRING = 9;</code>
-     */
-    TYPE_STRING(9),
-    /**
-     * <pre>
-     * Field type group. Proto2 syntax only, and deprecated.
-     * </pre>
-     *
-     * <code>TYPE_GROUP = 10;</code>
-     */
-    TYPE_GROUP(10),
-    /**
-     * <pre>
-     * Field type message.
-     * </pre>
-     *
-     * <code>TYPE_MESSAGE = 11;</code>
-     */
-    TYPE_MESSAGE(11),
-    /**
-     * <pre>
-     * Field type bytes.
-     * </pre>
-     *
-     * <code>TYPE_BYTES = 12;</code>
-     */
-    TYPE_BYTES(12),
-    /**
-     * <pre>
-     * Field type uint32.
-     * </pre>
-     *
-     * <code>TYPE_UINT32 = 13;</code>
-     */
-    TYPE_UINT32(13),
-    /**
-     * <pre>
-     * Field type enum.
-     * </pre>
-     *
-     * <code>TYPE_ENUM = 14;</code>
-     */
-    TYPE_ENUM(14),
-    /**
-     * <pre>
-     * Field type sfixed32.
-     * </pre>
-     *
-     * <code>TYPE_SFIXED32 = 15;</code>
-     */
-    TYPE_SFIXED32(15),
-    /**
-     * <pre>
-     * Field type sfixed64.
-     * </pre>
-     *
-     * <code>TYPE_SFIXED64 = 16;</code>
-     */
-    TYPE_SFIXED64(16),
-    /**
-     * <pre>
-     * Field type sint32.
-     * </pre>
-     *
-     * <code>TYPE_SINT32 = 17;</code>
-     */
-    TYPE_SINT32(17),
-    /**
-     * <pre>
-     * Field type sint64.
-     * </pre>
-     *
-     * <code>TYPE_SINT64 = 18;</code>
-     */
-    TYPE_SINT64(18),
-    UNRECOGNIZED(-1),
-    ;
-
-    /**
-     * <pre>
-     * Field type unknown.
-     * </pre>
-     *
-     * <code>TYPE_UNKNOWN = 0;</code>
-     */
-    public static final int TYPE_UNKNOWN_VALUE = 0;
-    /**
-     * <pre>
-     * Field type double.
-     * </pre>
-     *
-     * <code>TYPE_DOUBLE = 1;</code>
-     */
-    public static final int TYPE_DOUBLE_VALUE = 1;
-    /**
-     * <pre>
-     * Field type float.
-     * </pre>
-     *
-     * <code>TYPE_FLOAT = 2;</code>
-     */
-    public static final int TYPE_FLOAT_VALUE = 2;
-    /**
-     * <pre>
-     * Field type int64.
-     * </pre>
-     *
-     * <code>TYPE_INT64 = 3;</code>
-     */
-    public static final int TYPE_INT64_VALUE = 3;
-    /**
-     * <pre>
-     * Field type uint64.
-     * </pre>
-     *
-     * <code>TYPE_UINT64 = 4;</code>
-     */
-    public static final int TYPE_UINT64_VALUE = 4;
-    /**
-     * <pre>
-     * Field type int32.
-     * </pre>
-     *
-     * <code>TYPE_INT32 = 5;</code>
-     */
-    public static final int TYPE_INT32_VALUE = 5;
-    /**
-     * <pre>
-     * Field type fixed64.
-     * </pre>
-     *
-     * <code>TYPE_FIXED64 = 6;</code>
-     */
-    public static final int TYPE_FIXED64_VALUE = 6;
-    /**
-     * <pre>
-     * Field type fixed32.
-     * </pre>
-     *
-     * <code>TYPE_FIXED32 = 7;</code>
-     */
-    public static final int TYPE_FIXED32_VALUE = 7;
-    /**
-     * <pre>
-     * Field type bool.
-     * </pre>
-     *
-     * <code>TYPE_BOOL = 8;</code>
-     */
-    public static final int TYPE_BOOL_VALUE = 8;
-    /**
-     * <pre>
-     * Field type string.
-     * </pre>
-     *
-     * <code>TYPE_STRING = 9;</code>
-     */
-    public static final int TYPE_STRING_VALUE = 9;
-    /**
-     * <pre>
-     * Field type group. Proto2 syntax only, and deprecated.
-     * </pre>
-     *
-     * <code>TYPE_GROUP = 10;</code>
-     */
-    public static final int TYPE_GROUP_VALUE = 10;
-    /**
-     * <pre>
-     * Field type message.
-     * </pre>
-     *
-     * <code>TYPE_MESSAGE = 11;</code>
-     */
-    public static final int TYPE_MESSAGE_VALUE = 11;
-    /**
-     * <pre>
-     * Field type bytes.
-     * </pre>
-     *
-     * <code>TYPE_BYTES = 12;</code>
-     */
-    public static final int TYPE_BYTES_VALUE = 12;
-    /**
-     * <pre>
-     * Field type uint32.
-     * </pre>
-     *
-     * <code>TYPE_UINT32 = 13;</code>
-     */
-    public static final int TYPE_UINT32_VALUE = 13;
-    /**
-     * <pre>
-     * Field type enum.
-     * </pre>
-     *
-     * <code>TYPE_ENUM = 14;</code>
-     */
-    public static final int TYPE_ENUM_VALUE = 14;
-    /**
-     * <pre>
-     * Field type sfixed32.
-     * </pre>
-     *
-     * <code>TYPE_SFIXED32 = 15;</code>
-     */
-    public static final int TYPE_SFIXED32_VALUE = 15;
-    /**
-     * <pre>
-     * Field type sfixed64.
-     * </pre>
-     *
-     * <code>TYPE_SFIXED64 = 16;</code>
-     */
-    public static final int TYPE_SFIXED64_VALUE = 16;
-    /**
-     * <pre>
-     * Field type sint32.
-     * </pre>
-     *
-     * <code>TYPE_SINT32 = 17;</code>
-     */
-    public static final int TYPE_SINT32_VALUE = 17;
-    /**
-     * <pre>
-     * Field type sint64.
-     * </pre>
-     *
-     * <code>TYPE_SINT64 = 18;</code>
-     */
-    public static final int TYPE_SINT64_VALUE = 18;
-
-
-    public final int getNumber() {
-      if (this == UNRECOGNIZED) {
-        throw new java.lang.IllegalArgumentException(
-            "Can't get the number of an unknown enum value.");
-      }
-      return value;
-    }
-
-    /**
-     * @deprecated Use {@link #forNumber(int)} instead.
-     */
-    @java.lang.Deprecated
-    public static Kind valueOf(int value) {
-      return forNumber(value);
-    }
-
-    public static Kind forNumber(int value) {
-      switch (value) {
-        case 0: return TYPE_UNKNOWN;
-        case 1: return TYPE_DOUBLE;
-        case 2: return TYPE_FLOAT;
-        case 3: return TYPE_INT64;
-        case 4: return TYPE_UINT64;
-        case 5: return TYPE_INT32;
-        case 6: return TYPE_FIXED64;
-        case 7: return TYPE_FIXED32;
-        case 8: return TYPE_BOOL;
-        case 9: return TYPE_STRING;
-        case 10: return TYPE_GROUP;
-        case 11: return TYPE_MESSAGE;
-        case 12: return TYPE_BYTES;
-        case 13: return TYPE_UINT32;
-        case 14: return TYPE_ENUM;
-        case 15: return TYPE_SFIXED32;
-        case 16: return TYPE_SFIXED64;
-        case 17: return TYPE_SINT32;
-        case 18: return TYPE_SINT64;
-        default: return null;
-      }
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Kind>
-        internalGetValueMap() {
-      return internalValueMap;
-    }
-    private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
-        Kind> internalValueMap =
-          new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Kind>() {
-            public Kind findValueByNumber(int number) {
-              return Kind.forNumber(number);
-            }
-          };
-
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
-        getValueDescriptor() {
-      return getDescriptor().getValues().get(ordinal());
-    }
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.getDescriptor().getEnumTypes().get(0);
-    }
-
-    private static final Kind[] VALUES = values();
-
-    public static Kind valueOf(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-      if (desc.getType() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "EnumValueDescriptor is not for this type.");
-      }
-      if (desc.getIndex() == -1) {
-        return UNRECOGNIZED;
-      }
-      return VALUES[desc.getIndex()];
-    }
-
-    private final int value;
-
-    private Kind(int value) {
-      this.value = value;
-    }
-
-    // @@protoc_insertion_point(enum_scope:google.protobuf.Field.Kind)
-  }
-
-  /**
-   * <pre>
-   * Whether a field is optional, required, or repeated.
-   * </pre>
-   *
-   * Protobuf enum {@code google.protobuf.Field.Cardinality}
-   */
-  public enum Cardinality
-      implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
-    /**
-     * <pre>
-     * For fields with unknown cardinality.
-     * </pre>
-     *
-     * <code>CARDINALITY_UNKNOWN = 0;</code>
-     */
-    CARDINALITY_UNKNOWN(0),
-    /**
-     * <pre>
-     * For optional fields.
-     * </pre>
-     *
-     * <code>CARDINALITY_OPTIONAL = 1;</code>
-     */
-    CARDINALITY_OPTIONAL(1),
-    /**
-     * <pre>
-     * For required fields. Proto2 syntax only.
-     * </pre>
-     *
-     * <code>CARDINALITY_REQUIRED = 2;</code>
-     */
-    CARDINALITY_REQUIRED(2),
-    /**
-     * <pre>
-     * For repeated fields.
-     * </pre>
-     *
-     * <code>CARDINALITY_REPEATED = 3;</code>
-     */
-    CARDINALITY_REPEATED(3),
-    UNRECOGNIZED(-1),
-    ;
-
-    /**
-     * <pre>
-     * For fields with unknown cardinality.
-     * </pre>
-     *
-     * <code>CARDINALITY_UNKNOWN = 0;</code>
-     */
-    public static final int CARDINALITY_UNKNOWN_VALUE = 0;
-    /**
-     * <pre>
-     * For optional fields.
-     * </pre>
-     *
-     * <code>CARDINALITY_OPTIONAL = 1;</code>
-     */
-    public static final int CARDINALITY_OPTIONAL_VALUE = 1;
-    /**
-     * <pre>
-     * For required fields. Proto2 syntax only.
-     * </pre>
-     *
-     * <code>CARDINALITY_REQUIRED = 2;</code>
-     */
-    public static final int CARDINALITY_REQUIRED_VALUE = 2;
-    /**
-     * <pre>
-     * For repeated fields.
-     * </pre>
-     *
-     * <code>CARDINALITY_REPEATED = 3;</code>
-     */
-    public static final int CARDINALITY_REPEATED_VALUE = 3;
-
-
-    public final int getNumber() {
-      if (this == UNRECOGNIZED) {
-        throw new java.lang.IllegalArgumentException(
-            "Can't get the number of an unknown enum value.");
-      }
-      return value;
-    }
-
-    /**
-     * @deprecated Use {@link #forNumber(int)} instead.
-     */
-    @java.lang.Deprecated
-    public static Cardinality valueOf(int value) {
-      return forNumber(value);
-    }
-
-    public static Cardinality forNumber(int value) {
-      switch (value) {
-        case 0: return CARDINALITY_UNKNOWN;
-        case 1: return CARDINALITY_OPTIONAL;
-        case 2: return CARDINALITY_REQUIRED;
-        case 3: return CARDINALITY_REPEATED;
-        default: return null;
-      }
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Cardinality>
-        internalGetValueMap() {
-      return internalValueMap;
-    }
-    private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
-        Cardinality> internalValueMap =
-          new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Cardinality>() {
-            public Cardinality findValueByNumber(int number) {
-              return Cardinality.forNumber(number);
-            }
-          };
-
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
-        getValueDescriptor() {
-      return getDescriptor().getValues().get(ordinal());
-    }
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.getDescriptor().getEnumTypes().get(1);
-    }
-
-    private static final Cardinality[] VALUES = values();
-
-    public static Cardinality valueOf(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-      if (desc.getType() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "EnumValueDescriptor is not for this type.");
-      }
-      if (desc.getIndex() == -1) {
-        return UNRECOGNIZED;
-      }
-      return VALUES[desc.getIndex()];
-    }
-
-    private final int value;
-
-    private Cardinality(int value) {
-      this.value = value;
-    }
-
-    // @@protoc_insertion_point(enum_scope:google.protobuf.Field.Cardinality)
-  }
-
-  private int bitField0_;
-  public static final int KIND_FIELD_NUMBER = 1;
-  private int kind_;
-  /**
-   * <pre>
-   * The field type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Field.Kind kind = 1;</code>
-   */
-  public int getKindValue() {
-    return kind_;
-  }
-  /**
-   * <pre>
-   * The field type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Field.Kind kind = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind getKind() {
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind.valueOf(kind_);
-    return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind.UNRECOGNIZED : result;
-  }
-
-  public static final int CARDINALITY_FIELD_NUMBER = 2;
-  private int cardinality_;
-  /**
-   * <pre>
-   * The field cardinality.
-   * </pre>
-   *
-   * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-   */
-  public int getCardinalityValue() {
-    return cardinality_;
-  }
-  /**
-   * <pre>
-   * The field cardinality.
-   * </pre>
-   *
-   * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality getCardinality() {
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality.valueOf(cardinality_);
-    return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality.UNRECOGNIZED : result;
-  }
-
-  public static final int NUMBER_FIELD_NUMBER = 3;
-  private int number_;
-  /**
-   * <pre>
-   * The field number.
-   * </pre>
-   *
-   * <code>int32 number = 3;</code>
-   */
-  public int getNumber() {
-    return number_;
-  }
-
-  public static final int NAME_FIELD_NUMBER = 4;
-  private volatile java.lang.Object name_;
-  /**
-   * <pre>
-   * The field name.
-   * </pre>
-   *
-   * <code>string name = 4;</code>
-   */
-  public java.lang.String getName() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      name_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The field name.
-   * </pre>
-   *
-   * <code>string name = 4;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      name_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int TYPE_URL_FIELD_NUMBER = 6;
-  private volatile java.lang.Object typeUrl_;
-  /**
-   * <pre>
-   * The field type URL, without the scheme, for message or enumeration
-   * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-   * </pre>
-   *
-   * <code>string type_url = 6;</code>
-   */
-  public java.lang.String getTypeUrl() {
-    java.lang.Object ref = typeUrl_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      typeUrl_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The field type URL, without the scheme, for message or enumeration
-   * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-   * </pre>
-   *
-   * <code>string type_url = 6;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getTypeUrlBytes() {
-    java.lang.Object ref = typeUrl_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      typeUrl_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int ONEOF_INDEX_FIELD_NUMBER = 7;
-  private int oneofIndex_;
-  /**
-   * <pre>
-   * The index of the field type in `Type.oneofs`, for message or enumeration
-   * types. The first type has index 1; zero means the type is not in the list.
-   * </pre>
-   *
-   * <code>int32 oneof_index = 7;</code>
-   */
-  public int getOneofIndex() {
-    return oneofIndex_;
-  }
-
-  public static final int PACKED_FIELD_NUMBER = 8;
-  private boolean packed_;
-  /**
-   * <pre>
-   * Whether to use alternative packed wire representation.
-   * </pre>
-   *
-   * <code>bool packed = 8;</code>
-   */
-  public boolean getPacked() {
-    return packed_;
-  }
-
-  public static final int OPTIONS_FIELD_NUMBER = 9;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_;
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  public int getOptionsCount() {
-    return options_.size();
-  }
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-    return options_.get(index);
-  }
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index) {
-    return options_.get(index);
-  }
-
-  public static final int JSON_NAME_FIELD_NUMBER = 10;
-  private volatile java.lang.Object jsonName_;
-  /**
-   * <pre>
-   * The field JSON name.
-   * </pre>
-   *
-   * <code>string json_name = 10;</code>
-   */
-  public java.lang.String getJsonName() {
-    java.lang.Object ref = jsonName_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      jsonName_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The field JSON name.
-   * </pre>
-   *
-   * <code>string json_name = 10;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getJsonNameBytes() {
-    java.lang.Object ref = jsonName_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      jsonName_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int DEFAULT_VALUE_FIELD_NUMBER = 11;
-  private volatile java.lang.Object defaultValue_;
-  /**
-   * <pre>
-   * The string value of the default value of this field. Proto2 syntax only.
-   * </pre>
-   *
-   * <code>string default_value = 11;</code>
-   */
-  public java.lang.String getDefaultValue() {
-    java.lang.Object ref = defaultValue_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      defaultValue_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The string value of the default value of this field. Proto2 syntax only.
-   * </pre>
-   *
-   * <code>string default_value = 11;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getDefaultValueBytes() {
-    java.lang.Object ref = defaultValue_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      defaultValue_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (kind_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind.TYPE_UNKNOWN.getNumber()) {
-      output.writeEnum(1, kind_);
-    }
-    if (cardinality_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality.CARDINALITY_UNKNOWN.getNumber()) {
-      output.writeEnum(2, cardinality_);
-    }
-    if (number_ != 0) {
-      output.writeInt32(3, number_);
-    }
-    if (!getNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, name_);
-    }
-    if (!getTypeUrlBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 6, typeUrl_);
-    }
-    if (oneofIndex_ != 0) {
-      output.writeInt32(7, oneofIndex_);
-    }
-    if (packed_ != false) {
-      output.writeBool(8, packed_);
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      output.writeMessage(9, options_.get(i));
-    }
-    if (!getJsonNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 10, jsonName_);
-    }
-    if (!getDefaultValueBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 11, defaultValue_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (kind_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind.TYPE_UNKNOWN.getNumber()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeEnumSize(1, kind_);
-    }
-    if (cardinality_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality.CARDINALITY_UNKNOWN.getNumber()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeEnumSize(2, cardinality_);
-    }
-    if (number_ != 0) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt32Size(3, number_);
-    }
-    if (!getNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, name_);
-    }
-    if (!getTypeUrlBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(6, typeUrl_);
-    }
-    if (oneofIndex_ != 0) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt32Size(7, oneofIndex_);
-    }
-    if (packed_ != false) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeBoolSize(8, packed_);
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(9, options_.get(i));
-    }
-    if (!getJsonNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(10, jsonName_);
-    }
-    if (!getDefaultValueBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(11, defaultValue_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Field)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Field other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Field) obj;
-
-    boolean result = true;
-    result = result && kind_ == other.kind_;
-    result = result && cardinality_ == other.cardinality_;
-    result = result && (getNumber()
-        == other.getNumber());
-    result = result && getName()
-        .equals(other.getName());
-    result = result && getTypeUrl()
-        .equals(other.getTypeUrl());
-    result = result && (getOneofIndex()
-        == other.getOneofIndex());
-    result = result && (getPacked()
-        == other.getPacked());
-    result = result && getOptionsList()
-        .equals(other.getOptionsList());
-    result = result && getJsonName()
-        .equals(other.getJsonName());
-    result = result && getDefaultValue()
-        .equals(other.getDefaultValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + KIND_FIELD_NUMBER;
-    hash = (53 * hash) + kind_;
-    hash = (37 * hash) + CARDINALITY_FIELD_NUMBER;
-    hash = (53 * hash) + cardinality_;
-    hash = (37 * hash) + NUMBER_FIELD_NUMBER;
-    hash = (53 * hash) + getNumber();
-    hash = (37 * hash) + NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getName().hashCode();
-    hash = (37 * hash) + TYPE_URL_FIELD_NUMBER;
-    hash = (53 * hash) + getTypeUrl().hashCode();
-    hash = (37 * hash) + ONEOF_INDEX_FIELD_NUMBER;
-    hash = (53 * hash) + getOneofIndex();
-    hash = (37 * hash) + PACKED_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-        getPacked());
-    if (getOptionsCount() > 0) {
-      hash = (37 * hash) + OPTIONS_FIELD_NUMBER;
-      hash = (53 * hash) + getOptionsList().hashCode();
-    }
-    hash = (37 * hash) + JSON_NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getJsonName().hashCode();
-    hash = (37 * hash) + DEFAULT_VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + getDefaultValue().hashCode();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Field prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * A single field of a message type.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Field}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Field)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Field_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Field_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-        getOptionsFieldBuilder();
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      kind_ = 0;
-
-      cardinality_ = 0;
-
-      number_ = 0;
-
-      name_ = "";
-
-      typeUrl_ = "";
-
-      oneofIndex_ = 0;
-
-      packed_ = false;
-
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000080);
-      } else {
-        optionsBuilder_.clear();
-      }
-      jsonName_ = "";
-
-      defaultValue_ = "";
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Field_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Field result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Field result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Field(this);
-      int from_bitField0_ = bitField0_;
-      int to_bitField0_ = 0;
-      result.kind_ = kind_;
-      result.cardinality_ = cardinality_;
-      result.number_ = number_;
-      result.name_ = name_;
-      result.typeUrl_ = typeUrl_;
-      result.oneofIndex_ = oneofIndex_;
-      result.packed_ = packed_;
-      if (optionsBuilder_ == null) {
-        if (((bitField0_ & 0x00000080) == 0x00000080)) {
-          options_ = java.util.Collections.unmodifiableList(options_);
-          bitField0_ = (bitField0_ & ~0x00000080);
-        }
-        result.options_ = options_;
-      } else {
-        result.options_ = optionsBuilder_.build();
-      }
-      result.jsonName_ = jsonName_;
-      result.defaultValue_ = defaultValue_;
-      result.bitField0_ = to_bitField0_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Field) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Field)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Field other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.getDefaultInstance()) return this;
-      if (other.kind_ != 0) {
-        setKindValue(other.getKindValue());
-      }
-      if (other.cardinality_ != 0) {
-        setCardinalityValue(other.getCardinalityValue());
-      }
-      if (other.getNumber() != 0) {
-        setNumber(other.getNumber());
-      }
-      if (!other.getName().isEmpty()) {
-        name_ = other.name_;
-        onChanged();
-      }
-      if (!other.getTypeUrl().isEmpty()) {
-        typeUrl_ = other.typeUrl_;
-        onChanged();
-      }
-      if (other.getOneofIndex() != 0) {
-        setOneofIndex(other.getOneofIndex());
-      }
-      if (other.getPacked() != false) {
-        setPacked(other.getPacked());
-      }
-      if (optionsBuilder_ == null) {
-        if (!other.options_.isEmpty()) {
-          if (options_.isEmpty()) {
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000080);
-          } else {
-            ensureOptionsIsMutable();
-            options_.addAll(other.options_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.options_.isEmpty()) {
-          if (optionsBuilder_.isEmpty()) {
-            optionsBuilder_.dispose();
-            optionsBuilder_ = null;
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000080);
-            optionsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getOptionsFieldBuilder() : null;
-          } else {
-            optionsBuilder_.addAllMessages(other.options_);
-          }
-        }
-      }
-      if (!other.getJsonName().isEmpty()) {
-        jsonName_ = other.jsonName_;
-        onChanged();
-      }
-      if (!other.getDefaultValue().isEmpty()) {
-        defaultValue_ = other.defaultValue_;
-        onChanged();
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Field parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Field) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private int kind_ = 0;
-    /**
-     * <pre>
-     * The field type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Kind kind = 1;</code>
-     */
-    public int getKindValue() {
-      return kind_;
-    }
-    /**
-     * <pre>
-     * The field type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Kind kind = 1;</code>
-     */
-    public Builder setKindValue(int value) {
-      kind_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Kind kind = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind getKind() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind.valueOf(kind_);
-      return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind.UNRECOGNIZED : result;
-    }
-    /**
-     * <pre>
-     * The field type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Kind kind = 1;</code>
-     */
-    public Builder setKind(org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind value) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-      
-      kind_ = value.getNumber();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field type.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Kind kind = 1;</code>
-     */
-    public Builder clearKind() {
-      
-      kind_ = 0;
-      onChanged();
-      return this;
-    }
-
-    private int cardinality_ = 0;
-    /**
-     * <pre>
-     * The field cardinality.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-     */
-    public int getCardinalityValue() {
-      return cardinality_;
-    }
-    /**
-     * <pre>
-     * The field cardinality.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-     */
-    public Builder setCardinalityValue(int value) {
-      cardinality_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field cardinality.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality getCardinality() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality.valueOf(cardinality_);
-      return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality.UNRECOGNIZED : result;
-    }
-    /**
-     * <pre>
-     * The field cardinality.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-     */
-    public Builder setCardinality(org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality value) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-      
-      cardinality_ = value.getNumber();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field cardinality.
-     * </pre>
-     *
-     * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-     */
-    public Builder clearCardinality() {
-      
-      cardinality_ = 0;
-      onChanged();
-      return this;
-    }
-
-    private int number_ ;
-    /**
-     * <pre>
-     * The field number.
-     * </pre>
-     *
-     * <code>int32 number = 3;</code>
-     */
-    public int getNumber() {
-      return number_;
-    }
-    /**
-     * <pre>
-     * The field number.
-     * </pre>
-     *
-     * <code>int32 number = 3;</code>
-     */
-    public Builder setNumber(int value) {
-      
-      number_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field number.
-     * </pre>
-     *
-     * <code>int32 number = 3;</code>
-     */
-    public Builder clearNumber() {
-      
-      number_ = 0;
-      onChanged();
-      return this;
-    }
-
-    private java.lang.Object name_ = "";
-    /**
-     * <pre>
-     * The field name.
-     * </pre>
-     *
-     * <code>string name = 4;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        name_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The field name.
-     * </pre>
-     *
-     * <code>string name = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The field name.
-     * </pre>
-     *
-     * <code>string name = 4;</code>
-     */
-    public Builder setName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      name_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field name.
-     * </pre>
-     *
-     * <code>string name = 4;</code>
-     */
-    public Builder clearName() {
-      
-      name_ = getDefaultInstance().getName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field name.
-     * </pre>
-     *
-     * <code>string name = 4;</code>
-     */
-    public Builder setNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      name_ = value;
-      onChanged();
-      return this;
-    }
-
-    private java.lang.Object typeUrl_ = "";
-    /**
-     * <pre>
-     * The field type URL, without the scheme, for message or enumeration
-     * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-     * </pre>
-     *
-     * <code>string type_url = 6;</code>
-     */
-    public java.lang.String getTypeUrl() {
-      java.lang.Object ref = typeUrl_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        typeUrl_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The field type URL, without the scheme, for message or enumeration
-     * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-     * </pre>
-     *
-     * <code>string type_url = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getTypeUrlBytes() {
-      java.lang.Object ref = typeUrl_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        typeUrl_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The field type URL, without the scheme, for message or enumeration
-     * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-     * </pre>
-     *
-     * <code>string type_url = 6;</code>
-     */
-    public Builder setTypeUrl(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      typeUrl_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field type URL, without the scheme, for message or enumeration
-     * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-     * </pre>
-     *
-     * <code>string type_url = 6;</code>
-     */
-    public Builder clearTypeUrl() {
-      
-      typeUrl_ = getDefaultInstance().getTypeUrl();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field type URL, without the scheme, for message or enumeration
-     * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-     * </pre>
-     *
-     * <code>string type_url = 6;</code>
-     */
-    public Builder setTypeUrlBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      typeUrl_ = value;
-      onChanged();
-      return this;
-    }
-
-    private int oneofIndex_ ;
-    /**
-     * <pre>
-     * The index of the field type in `Type.oneofs`, for message or enumeration
-     * types. The first type has index 1; zero means the type is not in the list.
-     * </pre>
-     *
-     * <code>int32 oneof_index = 7;</code>
-     */
-    public int getOneofIndex() {
-      return oneofIndex_;
-    }
-    /**
-     * <pre>
-     * The index of the field type in `Type.oneofs`, for message or enumeration
-     * types. The first type has index 1; zero means the type is not in the list.
-     * </pre>
-     *
-     * <code>int32 oneof_index = 7;</code>
-     */
-    public Builder setOneofIndex(int value) {
-      
-      oneofIndex_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The index of the field type in `Type.oneofs`, for message or enumeration
-     * types. The first type has index 1; zero means the type is not in the list.
-     * </pre>
-     *
-     * <code>int32 oneof_index = 7;</code>
-     */
-    public Builder clearOneofIndex() {
-      
-      oneofIndex_ = 0;
-      onChanged();
-      return this;
-    }
-
-    private boolean packed_ ;
-    /**
-     * <pre>
-     * Whether to use alternative packed wire representation.
-     * </pre>
-     *
-     * <code>bool packed = 8;</code>
-     */
-    public boolean getPacked() {
-      return packed_;
-    }
-    /**
-     * <pre>
-     * Whether to use alternative packed wire representation.
-     * </pre>
-     *
-     * <code>bool packed = 8;</code>
-     */
-    public Builder setPacked(boolean value) {
-      
-      packed_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Whether to use alternative packed wire representation.
-     * </pre>
-     *
-     * <code>bool packed = 8;</code>
-     */
-    public Builder clearPacked() {
-      
-      packed_ = false;
-      onChanged();
-      return this;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_ =
-      java.util.Collections.emptyList();
-    private void ensureOptionsIsMutable() {
-      if (!((bitField0_ & 0x00000080) == 0x00000080)) {
-        options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>(options_);
-        bitField0_ |= 0x00000080;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> optionsBuilder_;
-
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-      if (optionsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(options_);
-      } else {
-        return optionsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public int getOptionsCount() {
-      if (optionsBuilder_ == null) {
-        return options_.size();
-      } else {
-        return optionsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);
-      } else {
-        return optionsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.set(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder addOptions(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder addOptions(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder addAllOptions(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> values) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, options_);
-        onChanged();
-      } else {
-        optionsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder clearOptions() {
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000080);
-        onChanged();
-      } else {
-        optionsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public Builder removeOptions(int index) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.remove(index);
-        onChanged();
-      } else {
-        optionsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder getOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-        int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);  } else {
-        return optionsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-         getOptionsOrBuilderList() {
-      if (optionsBuilder_ != null) {
-        return optionsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(options_);
-      }
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder() {
-      return getOptionsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * The protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 9;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder> 
-         getOptionsBuilderList() {
-      return getOptionsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-        getOptionsFieldBuilder() {
-      if (optionsBuilder_ == null) {
-        optionsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder>(
-                options_,
-                ((bitField0_ & 0x00000080) == 0x00000080),
-                getParentForChildren(),
-                isClean());
-        options_ = null;
-      }
-      return optionsBuilder_;
-    }
-
-    private java.lang.Object jsonName_ = "";
-    /**
-     * <pre>
-     * The field JSON name.
-     * </pre>
-     *
-     * <code>string json_name = 10;</code>
-     */
-    public java.lang.String getJsonName() {
-      java.lang.Object ref = jsonName_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        jsonName_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The field JSON name.
-     * </pre>
-     *
-     * <code>string json_name = 10;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getJsonNameBytes() {
-      java.lang.Object ref = jsonName_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        jsonName_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The field JSON name.
-     * </pre>
-     *
-     * <code>string json_name = 10;</code>
-     */
-    public Builder setJsonName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      jsonName_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field JSON name.
-     * </pre>
-     *
-     * <code>string json_name = 10;</code>
-     */
-    public Builder clearJsonName() {
-      
-      jsonName_ = getDefaultInstance().getJsonName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The field JSON name.
-     * </pre>
-     *
-     * <code>string json_name = 10;</code>
-     */
-    public Builder setJsonNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      jsonName_ = value;
-      onChanged();
-      return this;
-    }
-
-    private java.lang.Object defaultValue_ = "";
-    /**
-     * <pre>
-     * The string value of the default value of this field. Proto2 syntax only.
-     * </pre>
-     *
-     * <code>string default_value = 11;</code>
-     */
-    public java.lang.String getDefaultValue() {
-      java.lang.Object ref = defaultValue_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        defaultValue_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The string value of the default value of this field. Proto2 syntax only.
-     * </pre>
-     *
-     * <code>string default_value = 11;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getDefaultValueBytes() {
-      java.lang.Object ref = defaultValue_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        defaultValue_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The string value of the default value of this field. Proto2 syntax only.
-     * </pre>
-     *
-     * <code>string default_value = 11;</code>
-     */
-    public Builder setDefaultValue(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      defaultValue_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The string value of the default value of this field. Proto2 syntax only.
-     * </pre>
-     *
-     * <code>string default_value = 11;</code>
-     */
-    public Builder clearDefaultValue() {
-      
-      defaultValue_ = getDefaultInstance().getDefaultValue();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The string value of the default value of this field. Proto2 syntax only.
-     * </pre>
-     *
-     * <code>string default_value = 11;</code>
-     */
-    public Builder setDefaultValueBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      defaultValue_ = value;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Field)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Field)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Field DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Field();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Field getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Field>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Field>() {
-    public Field parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Field(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Field> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Field> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Field getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-


[34/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DescriptorProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DescriptorProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DescriptorProtos.java
deleted file mode 100644
index 0468e6c..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/DescriptorProtos.java
+++ /dev/null
@@ -1,39141 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/descriptor.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class DescriptorProtos {
-  private DescriptorProtos() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  public interface FileDescriptorSetOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:google.protobuf.FileDescriptorSet)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> 
-        getFileList();
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto getFile(int index);
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    int getFileCount();
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder> 
-        getFileOrBuilderList();
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder getFileOrBuilder(
-        int index);
-  }
-  /**
-   * <pre>
-   * The protocol compiler can output a FileDescriptorSet containing the .proto
-   * files it parses.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.FileDescriptorSet}
-   */
-  public  static final class FileDescriptorSet extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:google.protobuf.FileDescriptorSet)
-      FileDescriptorSetOrBuilder {
-    // Use FileDescriptorSet.newBuilder() to construct.
-    private FileDescriptorSet(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private FileDescriptorSet() {
-      file_ = java.util.Collections.emptyList();
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private FileDescriptorSet(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-                file_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto>();
-                mutable_bitField0_ |= 0x00000001;
-              }
-              file_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.PARSER, extensionRegistry));
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-          file_ = java.util.Collections.unmodifiableList(file_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.internal_static_google_protobuf_FileDescriptorSet_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.internal_static_google_protobuf_FileDescriptorSet_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet.Builder.class);
-    }
-
-    public static final int FILE_FIELD_NUMBER = 1;
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> file_;
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> getFileList() {
-      return file_;
-    }
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder> 
-        getFileOrBuilderList() {
-      return file_;
-    }
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    public int getFileCount() {
-      return file_.size();
-    }
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto getFile(int index) {
-      return file_.get(index);
-    }
-    /**
-     * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder getFileOrBuilder(
-        int index) {
-      return file_.get(index);
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      for (int i = 0; i < getFileCount(); i++) {
-        if (!getFile(i).isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      for (int i = 0; i < file_.size(); i++) {
-        output.writeMessage(1, file_.get(i));
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      for (int i = 0; i < file_.size(); i++) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, file_.get(i));
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet) obj;
-
-      boolean result = true;
-      result = result && getFileList()
-          .equals(other.getFileList());
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (getFileCount() > 0) {
-        hash = (37 * hash) + FILE_FIELD_NUMBER;
-        hash = (53 * hash) + getFileList().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * <pre>
-     * The protocol compiler can output a FileDescriptorSet containing the .proto
-     * files it parses.
-     * </pre>
-     *
-     * Protobuf type {@code google.protobuf.FileDescriptorSet}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:google.protobuf.FileDescriptorSet)
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSetOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.internal_static_google_protobuf_FileDescriptorSet_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.internal_static_google_protobuf_FileDescriptorSet_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-          getFileFieldBuilder();
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        if (fileBuilder_ == null) {
-          file_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000001);
-        } else {
-          fileBuilder_.clear();
-        }
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.internal_static_google_protobuf_FileDescriptorSet_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet build() {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet buildPartial() {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet(this);
-        int from_bitField0_ = bitField0_;
-        if (fileBuilder_ == null) {
-          if (((bitField0_ & 0x00000001) == 0x00000001)) {
-            file_ = java.util.Collections.unmodifiableList(file_);
-            bitField0_ = (bitField0_ & ~0x00000001);
-          }
-          result.file_ = file_;
-        } else {
-          result.file_ = fileBuilder_.build();
-        }
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet other) {
-        if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet.getDefaultInstance()) return this;
-        if (fileBuilder_ == null) {
-          if (!other.file_.isEmpty()) {
-            if (file_.isEmpty()) {
-              file_ = other.file_;
-              bitField0_ = (bitField0_ & ~0x00000001);
-            } else {
-              ensureFileIsMutable();
-              file_.addAll(other.file_);
-            }
-            onChanged();
-          }
-        } else {
-          if (!other.file_.isEmpty()) {
-            if (fileBuilder_.isEmpty()) {
-              fileBuilder_.dispose();
-              fileBuilder_ = null;
-              file_ = other.file_;
-              bitField0_ = (bitField0_ & ~0x00000001);
-              fileBuilder_ = 
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                   getFileFieldBuilder() : null;
-            } else {
-              fileBuilder_.addAllMessages(other.file_);
-            }
-          }
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        for (int i = 0; i < getFileCount(); i++) {
-          if (!getFile(i).isInitialized()) {
-            return false;
-          }
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> file_ =
-        java.util.Collections.emptyList();
-      private void ensureFileIsMutable() {
-        if (!((bitField0_ & 0x00000001) == 0x00000001)) {
-          file_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto>(file_);
-          bitField0_ |= 0x00000001;
-         }
-      }
-
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder> fileBuilder_;
-
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> getFileList() {
-        if (fileBuilder_ == null) {
-          return java.util.Collections.unmodifiableList(file_);
-        } else {
-          return fileBuilder_.getMessageList();
-        }
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public int getFileCount() {
-        if (fileBuilder_ == null) {
-          return file_.size();
-        } else {
-          return fileBuilder_.getCount();
-        }
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto getFile(int index) {
-        if (fileBuilder_ == null) {
-          return file_.get(index);
-        } else {
-          return fileBuilder_.getMessage(index);
-        }
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder setFile(
-          int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto value) {
-        if (fileBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureFileIsMutable();
-          file_.set(index, value);
-          onChanged();
-        } else {
-          fileBuilder_.setMessage(index, value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder setFile(
-          int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder builderForValue) {
-        if (fileBuilder_ == null) {
-          ensureFileIsMutable();
-          file_.set(index, builderForValue.build());
-          onChanged();
-        } else {
-          fileBuilder_.setMessage(index, builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder addFile(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto value) {
-        if (fileBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureFileIsMutable();
-          file_.add(value);
-          onChanged();
-        } else {
-          fileBuilder_.addMessage(value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder addFile(
-          int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto value) {
-        if (fileBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          ensureFileIsMutable();
-          file_.add(index, value);
-          onChanged();
-        } else {
-          fileBuilder_.addMessage(index, value);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder addFile(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder builderForValue) {
-        if (fileBuilder_ == null) {
-          ensureFileIsMutable();
-          file_.add(builderForValue.build());
-          onChanged();
-        } else {
-          fileBuilder_.addMessage(builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder addFile(
-          int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder builderForValue) {
-        if (fileBuilder_ == null) {
-          ensureFileIsMutable();
-          file_.add(index, builderForValue.build());
-          onChanged();
-        } else {
-          fileBuilder_.addMessage(index, builderForValue.build());
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder addAllFile(
-          java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> values) {
-        if (fileBuilder_ == null) {
-          ensureFileIsMutable();
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-              values, file_);
-          onChanged();
-        } else {
-          fileBuilder_.addAllMessages(values);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder clearFile() {
-        if (fileBuilder_ == null) {
-          file_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000001);
-          onChanged();
-        } else {
-          fileBuilder_.clear();
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public Builder removeFile(int index) {
-        if (fileBuilder_ == null) {
-          ensureFileIsMutable();
-          file_.remove(index);
-          onChanged();
-        } else {
-          fileBuilder_.remove(index);
-        }
-        return this;
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder getFileBuilder(
-          int index) {
-        return getFileFieldBuilder().getBuilder(index);
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder getFileOrBuilder(
-          int index) {
-        if (fileBuilder_ == null) {
-          return file_.get(index);  } else {
-          return fileBuilder_.getMessageOrBuilder(index);
-        }
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder> 
-           getFileOrBuilderList() {
-        if (fileBuilder_ != null) {
-          return fileBuilder_.getMessageOrBuilderList();
-        } else {
-          return java.util.Collections.unmodifiableList(file_);
-        }
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder addFileBuilder() {
-        return getFileFieldBuilder().addBuilder(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.getDefaultInstance());
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder addFileBuilder(
-          int index) {
-        return getFileFieldBuilder().addBuilder(
-            index, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.getDefaultInstance());
-      }
-      /**
-       * <code>repeated .google.protobuf.FileDescriptorProto file = 1;</code>
-       */
-      public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder> 
-           getFileBuilderList() {
-        return getFileFieldBuilder().getBuilderList();
-      }
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder> 
-          getFileFieldBuilder() {
-        if (fileBuilder_ == null) {
-          fileBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder>(
-                  file_,
-                  ((bitField0_ & 0x00000001) == 0x00000001),
-                  getParentForChildren(),
-                  isClean());
-          file_ = null;
-        }
-        return fileBuilder_;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:google.protobuf.FileDescriptorSet)
-    }
-
-    // @@protoc_insertion_point(class_scope:google.protobuf.FileDescriptorSet)
-    private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileDescriptorSet>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<FileDescriptorSet>() {
-      public FileDescriptorSet parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new FileDescriptorSet(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileDescriptorSet> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FileDescriptorSet> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorSet getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface FileDescriptorProtoOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:google.protobuf.FileDescriptorProto)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <pre>
-     * file name, relative to root of source tree
-     * </pre>
-     *
-     * <code>optional string name = 1;</code>
-     */
-    boolean hasName();
-    /**
-     * <pre>
-     * file name, relative to root of source tree
-     * </pre>
-     *
-     * <code>optional string name = 1;</code>
-     */
-    java.lang.String getName();
-    /**
-     * <pre>
-     * file name, relative to root of source tree
-     * </pre>
-     *
-     * <code>optional string name = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes();
-
-    /**
-     * <pre>
-     * e.g. "foo", "foo.bar", etc.
-     * </pre>
-     *
-     * <code>optional string package = 2;</code>
-     */
-    boolean hasPackage();
-    /**
-     * <pre>
-     * e.g. "foo", "foo.bar", etc.
-     * </pre>
-     *
-     * <code>optional string package = 2;</code>
-     */
-    java.lang.String getPackage();
-    /**
-     * <pre>
-     * e.g. "foo", "foo.bar", etc.
-     * </pre>
-     *
-     * <code>optional string package = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getPackageBytes();
-
-    /**
-     * <pre>
-     * Names of files imported by this file.
-     * </pre>
-     *
-     * <code>repeated string dependency = 3;</code>
-     */
-    java.util.List<java.lang.String>
-        getDependencyList();
-    /**
-     * <pre>
-     * Names of files imported by this file.
-     * </pre>
-     *
-     * <code>repeated string dependency = 3;</code>
-     */
-    int getDependencyCount();
-    /**
-     * <pre>
-     * Names of files imported by this file.
-     * </pre>
-     *
-     * <code>repeated string dependency = 3;</code>
-     */
-    java.lang.String getDependency(int index);
-    /**
-     * <pre>
-     * Names of files imported by this file.
-     * </pre>
-     *
-     * <code>repeated string dependency = 3;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getDependencyBytes(int index);
-
-    /**
-     * <pre>
-     * Indexes of the public imported files in the dependency list above.
-     * </pre>
-     *
-     * <code>repeated int32 public_dependency = 10;</code>
-     */
-    java.util.List<java.lang.Integer> getPublicDependencyList();
-    /**
-     * <pre>
-     * Indexes of the public imported files in the dependency list above.
-     * </pre>
-     *
-     * <code>repeated int32 public_dependency = 10;</code>
-     */
-    int getPublicDependencyCount();
-    /**
-     * <pre>
-     * Indexes of the public imported files in the dependency list above.
-     * </pre>
-     *
-     * <code>repeated int32 public_dependency = 10;</code>
-     */
-    int getPublicDependency(int index);
-
-    /**
-     * <pre>
-     * Indexes of the weak imported files in the dependency list.
-     * For Google-internal migration only. Do not use.
-     * </pre>
-     *
-     * <code>repeated int32 weak_dependency = 11;</code>
-     */
-    java.util.List<java.lang.Integer> getWeakDependencyList();
-    /**
-     * <pre>
-     * Indexes of the weak imported files in the dependency list.
-     * For Google-internal migration only. Do not use.
-     * </pre>
-     *
-     * <code>repeated int32 weak_dependency = 11;</code>
-     */
-    int getWeakDependencyCount();
-    /**
-     * <pre>
-     * Indexes of the weak imported files in the dependency list.
-     * For Google-internal migration only. Do not use.
-     * </pre>
-     *
-     * <code>repeated int32 weak_dependency = 11;</code>
-     */
-    int getWeakDependency(int index);
-
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProto> 
-        getMessageTypeList();
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProto getMessageType(int index);
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    int getMessageTypeCount();
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProtoOrBuilder> 
-        getMessageTypeOrBuilderList();
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProtoOrBuilder getMessageTypeOrBuilder(
-        int index);
-
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProto> 
-        getEnumTypeList();
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProto getEnumType(int index);
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    int getEnumTypeCount();
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProtoOrBuilder> 
-        getEnumTypeOrBuilderList();
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProtoOrBuilder getEnumTypeOrBuilder(
-        int index);
-
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProto> 
-        getServiceList();
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProto getService(int index);
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    int getServiceCount();
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProtoOrBuilder> 
-        getServiceOrBuilderList();
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProtoOrBuilder getServiceOrBuilder(
-        int index);
-
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProto> 
-        getExtensionList();
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProto getExtension(int index);
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    int getExtensionCount();
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProtoOrBuilder> 
-        getExtensionOrBuilderList();
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProtoOrBuilder getExtensionOrBuilder(
-        int index);
-
-    /**
-     * <code>optional .google.protobuf.FileOptions options = 8;</code>
-     */
-    boolean hasOptions();
-    /**
-     * <code>optional .google.protobuf.FileOptions options = 8;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptions getOptions();
-    /**
-     * <code>optional .google.protobuf.FileOptions options = 8;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptionsOrBuilder getOptionsOrBuilder();
-
-    /**
-     * <pre>
-     * This field contains optional information about the original source code.
-     * You may safely remove this entire field without harming runtime
-     * functionality of the descriptors -- the information is needed only by
-     * development tools.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.SourceCodeInfo source_code_info = 9;</code>
-     */
-    boolean hasSourceCodeInfo();
-    /**
-     * <pre>
-     * This field contains optional information about the original source code.
-     * You may safely remove this entire field without harming runtime
-     * functionality of the descriptors -- the information is needed only by
-     * development tools.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.SourceCodeInfo source_code_info = 9;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfo getSourceCodeInfo();
-    /**
-     * <pre>
-     * This field contains optional information about the original source code.
-     * You may safely remove this entire field without harming runtime
-     * functionality of the descriptors -- the information is needed only by
-     * development tools.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.SourceCodeInfo source_code_info = 9;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfoOrBuilder getSourceCodeInfoOrBuilder();
-
-    /**
-     * <pre>
-     * The syntax of the proto file.
-     * The supported values are "proto2" and "proto3".
-     * </pre>
-     *
-     * <code>optional string syntax = 12;</code>
-     */
-    boolean hasSyntax();
-    /**
-     * <pre>
-     * The syntax of the proto file.
-     * The supported values are "proto2" and "proto3".
-     * </pre>
-     *
-     * <code>optional string syntax = 12;</code>
-     */
-    java.lang.String getSyntax();
-    /**
-     * <pre>
-     * The syntax of the proto file.
-     * The supported values are "proto2" and "proto3".
-     * </pre>
-     *
-     * <code>optional string syntax = 12;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getSyntaxBytes();
-  }
-  /**
-   * <pre>
-   * Describes a complete .proto file.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.FileDescriptorProto}
-   */
-  public  static final class FileDescriptorProto extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:google.protobuf.FileDescriptorProto)
-      FileDescriptorProtoOrBuilder {
-    // Use FileDescriptorProto.newBuilder() to construct.
-    private FileDescriptorProto(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private FileDescriptorProto() {
-      name_ = "";
-      package_ = "";
-      dependency_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-      publicDependency_ = java.util.Collections.emptyList();
-      weakDependency_ = java.util.Collections.emptyList();
-      messageType_ = java.util.Collections.emptyList();
-      enumType_ = java.util.Collections.emptyList();
-      service_ = java.util.Collections.emptyList();
-      extension_ = java.util.Collections.emptyList();
-      syntax_ = "";
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private FileDescriptorProto(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000001;
-              name_ = bs;
-              break;
-            }
-            case 18: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000002;
-              package_ = bs;
-              break;
-            }
-            case 26: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-                dependency_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList();
-                mutable_bitField0_ |= 0x00000004;
-              }
-              dependency_.add(bs);
-              break;
-            }
-            case 34: {
-              if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
-                messageType_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProto>();
-                mutable_bitField0_ |= 0x00000020;
-              }
-              messageType_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProto.PARSER, extensionRegistry));
-              break;
-            }
-            case 42: {
-              if (!((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
-                enumType_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProto>();
-                mutable_bitField0_ |= 0x00000040;
-              }
-              enumType_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProto.PARSER, extensionRegistry));
-              break;
-            }
-            case 50: {
-              if (!((mutable_bitField0_ & 0x00000080) == 0x00000080)) {
-                service_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProto>();
-                mutable_bitField0_ |= 0x00000080;
-              }
-              service_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProto.PARSER, extensionRegistry));
-              break;
-            }
-            case 58: {
-              if (!((mutable_bitField0_ & 0x00000100) == 0x00000100)) {
-                extension_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProto>();
-                mutable_bitField0_ |= 0x00000100;
-              }
-              extension_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProto.PARSER, extensionRegistry));
-              break;
-            }
-            case 66: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptions.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000004) == 0x00000004)) {
-                subBuilder = options_.toBuilder();
-              }
-              options_ = input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptions.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(options_);
-                options_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000004;
-              break;
-            }
-            case 74: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfo.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000008) == 0x00000008)) {
-                subBuilder = sourceCodeInfo_.toBuilder();
-              }
-              sourceCodeInfo_ = input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfo.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(sourceCodeInfo_);
-                sourceCodeInfo_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000008;
-              break;
-            }
-            case 80: {
-              if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-                publicDependency_ = new java.util.ArrayList<java.lang.Integer>();
-                mutable_bitField0_ |= 0x00000008;
-              }
-              publicDependency_.add(input.readInt32());
-              break;
-            }
-            case 82: {
-              int length = input.readRawVarint32();
-              int limit = input.pushLimit(length);
-              if (!((mutable_bitField0_ & 0x00000008) == 0x00000008) && input.getBytesUntilLimit() > 0) {
-                publicDependency_ = new java.util.ArrayList<java.lang.Integer>();
-                mutable_bitField0_ |= 0x00000008;
-              }
-              while (input.getBytesUntilLimit() > 0) {
-                publicDependency_.add(input.readInt32());
-              }
-              input.popLimit(limit);
-              break;
-            }
-            case 88: {
-              if (!((mutable_bitField0_ & 0x00000010) == 0x00000010)) {
-                weakDependency_ = new java.util.ArrayList<java.lang.Integer>();
-                mutable_bitField0_ |= 0x00000010;
-              }
-              weakDependency_.add(input.readInt32());
-              break;
-            }
-            case 90: {
-              int length = input.readRawVarint32();
-              int limit = input.pushLimit(length);
-              if (!((mutable_bitField0_ & 0x00000010) == 0x00000010) && input.getBytesUntilLimit() > 0) {
-                weakDependency_ = new java.util.ArrayList<java.lang.Integer>();
-                mutable_bitField0_ |= 0x00000010;
-              }
-              while (input.getBytesUntilLimit() > 0) {
-                weakDependency_.add(input.readInt32());
-              }
-              input.popLimit(limit);
-              break;
-            }
-            case 98: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000010;
-              syntax_ = bs;
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-          dependency_ = dependency_.getUnmodifiableView();
-        }
-        if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
-          messageType_ = java.util.Collections.unmodifiableList(messageType_);
-        }
-        if (((mutable_bitField0_ & 0x00000040) == 0x00000040)) {
-          enumType_ = java.util.Collections.unmodifiableList(enumType_);
-        }
-        if (((mutable_bitField0_ & 0x00000080) == 0x00000080)) {
-          service_ = java.util.Collections.unmodifiableList(service_);
-        }
-        if (((mutable_bitField0_ & 0x00000100) == 0x00000100)) {
-          extension_ = java.util.Collections.unmodifiableList(extension_);
-        }
-        if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-          publicDependency_ = java.util.Collections.unmodifiableList(publicDependency_);
-        }
-        if (((mutable_bitField0_ & 0x00000010) == 0x00000010)) {
-          weakDependency_ = java.util.Collections.unmodifiableList(weakDependency_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.internal_static_google_protobuf_FileDescriptorProto_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.internal_static_google_protobuf_FileDescriptorProto_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int NAME_FIELD_NUMBER = 1;
-    private volatile java.lang.Object name_;
-    /**
-     * <pre>
-     * file name, relative to root of source tree
-     * </pre>
-     *
-     * <code>optional string name = 1;</code>
-     */
-    public boolean hasName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <pre>
-     * file name, relative to root of source tree
-     * </pre>
-     *
-     * <code>optional string name = 1;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          name_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <pre>
-     * file name, relative to root of source tree
-     * </pre>
-     *
-     * <code>optional string name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof java.lang.String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    public static final int PACKAGE_FIELD_NUMBER = 2;
-    private volatile java.lang.Object package_;
-    /**
-     * <pre>
-     * e.g. "foo", "foo.bar", etc.
-     * </pre>
-     *
-     * <code>optional string package = 2;</code>
-     */
-    public boolean hasPackage() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <pre>
-     * e.g. "foo", "foo.bar", etc.
-     * </pre>
-     *
-     * <code>optional string package = 2;</code>
-     */
-    public java.lang.String getPackage() {
-      java.lang.Object ref = package_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          package_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <pre>
-     * e.g. "foo", "foo.bar", etc.
-     * </pre>
-     *
-     * <code>optional string package = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getPackageBytes() {
-      java.lang.Object ref = package_;
-      if (ref instanceof java.lang.String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        package_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    public static final int DEPENDENCY_FIELD_NUMBER = 3;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList dependency_;
-    /**
-     * <pre>
-     * Names of files imported by this file.
-     * </pre>
-     *
-     * <code>repeated string dependency = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
-        getDependencyList() {
-      return dependency_;
-    }
-    /**
-     * <pre>
-     * Names of files imported by this file.
-     * </pre>
-     *
-     * <code>repeated string dependency = 3;</code>
-     */
-    public int getDependencyCount() {
-      return dependency_.size();
-    }
-    /**
-     * <pre>
-     * Names of files imported by this file.
-     * </pre>
-     *
-     * <code>repeated string dependency = 3;</code>
-     */
-    public java.lang.String getDependency(int index) {
-      return dependency_.get(index);
-    }
-    /**
-     * <pre>
-     * Names of files imported by this file.
-     * </pre>
-     *
-     * <code>repeated string dependency = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getDependencyBytes(int index) {
-      return dependency_.getByteString(index);
-    }
-
-    public static final int PUBLIC_DEPENDENCY_FIELD_NUMBER = 10;
-    private java.util.List<java.lang.Integer> publicDependency_;
-    /**
-     * <pre>
-     * Indexes of the public imported files in the dependency list above.
-     * </pre>
-     *
-     * <code>repeated int32 public_dependency = 10;</code>
-     */
-    public java.util.List<java.lang.Integer>
-        getPublicDependencyList() {
-      return publicDependency_;
-    }
-    /**
-     * <pre>
-     * Indexes of the public imported files in the dependency list above.
-     * </pre>
-     *
-     * <code>repeated int32 public_dependency = 10;</code>
-     */
-    public int getPublicDependencyCount() {
-      return publicDependency_.size();
-    }
-    /**
-     * <pre>
-     * Indexes of the public imported files in the dependency list above.
-     * </pre>
-     *
-     * <code>repeated int32 public_dependency = 10;</code>
-     */
-    public int getPublicDependency(int index) {
-      return publicDependency_.get(index);
-    }
-
-    public static final int WEAK_DEPENDENCY_FIELD_NUMBER = 11;
-    private java.util.List<java.lang.Integer> weakDependency_;
-    /**
-     * <pre>
-     * Indexes of the weak imported files in the dependency list.
-     * For Google-internal migration only. Do not use.
-     * </pre>
-     *
-     * <code>repeated int32 weak_dependency = 11;</code>
-     */
-    public java.util.List<java.lang.Integer>
-        getWeakDependencyList() {
-      return weakDependency_;
-    }
-    /**
-     * <pre>
-     * Indexes of the weak imported files in the dependency list.
-     * For Google-internal migration only. Do not use.
-     * </pre>
-     *
-     * <code>repeated int32 weak_dependency = 11;</code>
-     */
-    public int getWeakDependencyCount() {
-      return weakDependency_.size();
-    }
-    /**
-     * <pre>
-     * Indexes of the weak imported files in the dependency list.
-     * For Google-internal migration only. Do not use.
-     * </pre>
-     *
-     * <code>repeated int32 weak_dependency = 11;</code>
-     */
-    public int getWeakDependency(int index) {
-      return weakDependency_.get(index);
-    }
-
-    public static final int MESSAGE_TYPE_FIELD_NUMBER = 4;
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProto> messageType_;
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProto> getMessageTypeList() {
-      return messageType_;
-    }
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProtoOrBuilder> 
-        getMessageTypeOrBuilderList() {
-      return messageType_;
-    }
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    public int getMessageTypeCount() {
-      return messageType_.size();
-    }
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProto getMessageType(int index) {
-      return messageType_.get(index);
-    }
-    /**
-     * <pre>
-     * All top-level definitions in this file.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.DescriptorProto message_type = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.DescriptorProtoOrBuilder getMessageTypeOrBuilder(
-        int index) {
-      return messageType_.get(index);
-    }
-
-    public static final int ENUM_TYPE_FIELD_NUMBER = 5;
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProto> enumType_;
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProto> getEnumTypeList() {
-      return enumType_;
-    }
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProtoOrBuilder> 
-        getEnumTypeOrBuilderList() {
-      return enumType_;
-    }
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    public int getEnumTypeCount() {
-      return enumType_.size();
-    }
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProto getEnumType(int index) {
-      return enumType_.get(index);
-    }
-    /**
-     * <code>repeated .google.protobuf.EnumDescriptorProto enum_type = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.EnumDescriptorProtoOrBuilder getEnumTypeOrBuilder(
-        int index) {
-      return enumType_.get(index);
-    }
-
-    public static final int SERVICE_FIELD_NUMBER = 6;
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProto> service_;
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProto> getServiceList() {
-      return service_;
-    }
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProtoOrBuilder> 
-        getServiceOrBuilderList() {
-      return service_;
-    }
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    public int getServiceCount() {
-      return service_.size();
-    }
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProto getService(int index) {
-      return service_.get(index);
-    }
-    /**
-     * <code>repeated .google.protobuf.ServiceDescriptorProto service = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.ServiceDescriptorProtoOrBuilder getServiceOrBuilder(
-        int index) {
-      return service_.get(index);
-    }
-
-    public static final int EXTENSION_FIELD_NUMBER = 7;
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProto> extension_;
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProto> getExtensionList() {
-      return extension_;
-    }
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProtoOrBuilder> 
-        getExtensionOrBuilderList() {
-      return extension_;
-    }
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    public int getExtensionCount() {
-      return extension_.size();
-    }
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProto getExtension(int index) {
-      return extension_.get(index);
-    }
-    /**
-     * <code>repeated .google.protobuf.FieldDescriptorProto extension = 7;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FieldDescriptorProtoOrBuilder getExtensionOrBuilder(
-        int index) {
-      return extension_.get(index);
-    }
-
-    public static final int OPTIONS_FIELD_NUMBER = 8;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptions options_;
-    /**
-     * <code>optional .google.protobuf.FileOptions options = 8;</code>
-     */
-    public boolean hasOptions() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>optional .google.protobuf.FileOptions options = 8;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptions getOptions() {
-      return options_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptions.getDefaultInstance() : options_;
-    }
-    /**
-     * <code>optional .google.protobuf.FileOptions options = 8;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptionsOrBuilder getOptionsOrBuilder() {
-      return options_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileOptions.getDefaultInstance() : options_;
-    }
-
-    public static final int SOURCE_CODE_INFO_FIELD_NUMBER = 9;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfo sourceCodeInfo_;
-    /**
-     * <pre>
-     * This field contains optional information about the original source code.
-     * You may safely remove this entire field without harming runtime
-     * functionality of the descriptors -- the information is needed only by
-     * development tools.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.SourceCodeInfo source_code_info = 9;</code>
-     */
-    public boolean hasSourceCodeInfo() {
-      return ((bitField0_ & 0x00000008) == 0x00000008);
-    }
-    /**
-     * <pre>
-     * This field contains optional information about the original source code.
-     * You may safely remove this entire field without harming runtime
-     * functionality of the descriptors -- the information is needed only by
-     * development tools.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.SourceCodeInfo source_code_info = 9;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfo getSourceCodeInfo() {
-      return sourceCodeInfo_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfo.getDefaultInstance() : sourceCodeInfo_;
-    }
-    /**
-     * <pre>
-     * This field contains optional information about the original source code.
-     * You may safely remove this entire field without harming runtime
-     * functionality of the descriptors -- the information is needed only by
-     * development tools.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.SourceCodeInfo source_code_info = 9;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfoOrBuilder getSourceCodeInfoOrBuilder() {
-      return sourceCodeInfo_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.SourceCodeInfo.getDefaultInstance() : sourceCodeInfo_;
-    }
-
-    public static final int SYNTAX_FIELD_NUMBER = 12;
-    private volatile java.lang.Object syntax_;
-    /**
-     * <pre>
-     * The syntax of the proto file.
-     * The supported values are "proto2" and "proto3".
-     * </pre>
-     *
-     * <code>optional string syntax = 12;</code>
-     */
-    public boolean hasSyntax() {
-      return ((bitField0_ & 0x00000010) == 0x00000010);
-    }
-    /**
-     * <pre>
-     * The syntax of the proto file.
-     * The supported values are "proto2" and "proto3".
-     * </pre>
-     *
-     * <code>optional string syntax = 12;</code>
-     */
-    public java.lang.String getSyntax() {
-      java.lang.Object ref = syntax_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          syntax_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <pre>
-     * The syntax of the proto file.
-     * The supported values are "proto2" and "proto3".
-     * </pre>
-     *
-     * <code>optional string syntax = 12;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getSyntaxBytes() {
-      java.lang.Object ref = syntax_;
-      if (ref instanceof java.lang.String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        syntax_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      for (int i = 0; i < getMessageTypeCount(); i++) {
-        if (!getMessageType(i).isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      for (int i = 0; i < getEnumTypeCount(); i++) {
-        if (!getEnumType(i).isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      for (int i = 0; i < getServiceCount(); i++) {
-        if (!getService(i).isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      for (int i = 0; i < getExtensionCount(); i++) {
-        if (!getExtension(i).isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      if (hasOptions()) {
-        if (!getOptions().isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, package_);
-      }
-      for (int i = 0; i < dependency_.size(); i++) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, dependency_.getRaw(i));
-      }
-      for (int i = 0; i < messageType_.size(); i++) {
-        output.writeMessage(4, messageType_.get(i));
-      }
-      for (int i = 0; i < enumType_.size(); i++) {
-        output.writeMessage(5, enumType_.get(i));
-      }
-      for (int i = 0; i < service_.size(); i++) {
-        output.writeMessage(6, service_.get(i));
-      }
-      for (int i = 0; i < extension_.size(); i++) {
-        output.writeMessage(7, extension_.get(i));
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeMessage(8, getOptions());
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        output.writeMessage(9, getSourceCodeInfo());
-      }
-      for (int i = 0; i < publicDependency_.size(); i++) {
-        output.writeInt32(10, publicDependency_.get(i));
-      }
-      for (int i = 0; i < weakDependency_.size(); i++) {
-        output.writeInt32(11, weakDependency_.get(i));
-      }
-      if (((bitField0_ & 0x00000010) == 0x00000010)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 12, syntax_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, package_);
-      }
-      {
-        int dataSize = 0;
-        for (int i = 0; i < dependency_.size(); i++) {
-          dataSize += computeStringSizeNoTag(dependency_.getRaw(i));
-        }
-        size += dataSize;
-        size += 1 * getDependencyList().size();
-      }
-      for (int i = 0; i < messageType_.size(); i++) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(4, messageType_.get(i));
-      }
-      for (int i = 0; i < enumType_.size(); i++) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(5, enumType_.get(i));
-      }
-      for (int i = 0; i < service_.size(); i++) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(6, service_.get(i));
-      }
-      for (int i = 0; i < extension_.size(); i++) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(7, extension_.get(i));
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(8, getOptions());
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(9, getSourceCodeInfo());
-      }
-      {
-        int dataSize = 0;
-        for (int i = 0; i < publicDependency_.size(); i++) {
-          dataSize += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-            .computeInt32SizeNoTag(publicDependency_.get(i));
-        }
-        size += dataSize;
-        size += 1 * getPublicDependencyList().size();
-      }
-      {
-        int dataSize = 0;
-        for (int i = 0; i < weakDependency_.size(); i++) {
-          dataSize += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-            .computeInt32SizeNoTag(weakDependency_.get(i));
-        }
-        size += dataSize;
-        size += 1 * getWeakDependencyList().size();
-      }
-      if (((bitField0_ & 0x00000010) == 0x00000010)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(12, syntax_);
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto) obj;
-
-      boolean result = true;
-      result = result && (hasName() == other.hasName());
-      if (hasName()) {
-        result = result && getName()
-            .equals(other.getName());
-      }
-      result = result && (hasPackage() == other.hasPackage());
-      if (hasPackage()) {
-        result = result && getPackage()
-            .equals(other.getPackage());
-      }
-      result = result && getDependencyList()
-          .equals(other.getDependencyList());
-      result = result && getPublicDependencyList()
-          .equals(other.getPublicDependencyList());
-      result = result && getWeakDependencyList()
-          .equals(other.getWeakDependencyList());
-      result = result && getMessageTypeList()
-          .equals(other.getMessageTypeList());
-      result = result && getEnumTypeList()
-          .equals(other.getEnumTypeList());
-      result = result && getServiceList()
-          .equals(other.getServiceList());
-      result = result && getExtensionList()
-          .equals(other.getExtensionList());
-      result = result && (hasOptions() == other.hasOptions());
-      if (hasOptions()) {
-        result = result && getOptions()
-            .equals(other.getOptions());
-      }
-      result = result && (hasSourceCodeInfo() == other.hasSourceCodeInfo());
-      if (hasSourceCodeInfo()) {
-        result = result && getSourceCodeInfo()
-            .equals(other.getSourceCodeInfo());
-      }
-      result = result && (hasSyntax() == other.hasSyntax());
-      if (hasSyntax()) {
-        result = result && getSyntax()
-            .equals(other.getSyntax());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasName()) {
-        hash = (37 * hash) + NAME_FIELD_NUMBER;
-        hash = (53 * hash) + getName().hashCode();
-      }
-      if (hasPackage()) {
-        hash = (37 * hash) + PACKAGE_FIELD_NUMBER;
-        hash = (53 * hash) + getPackage().hashCode();
-      }
-      if (getDependencyCount() > 0) {
-        hash = (37 * hash) + DEPENDENCY_FIELD_NUMBER;
-        hash = (53 * hash) + getDependencyList().hashCode();
-      }
-      if (getPublicDependencyCount() > 0) {
-        hash = (37 * hash) + PUBLIC_DEPENDENCY_FIELD_NUMBER;
-        hash = (53 * hash) + getPublicDependencyList().hashCode();
-      }
-      if (getWeakDependencyCount() > 0) {
-        hash = (37 * hash) + WEAK_DEPENDENCY_FIELD_NUMBER;
-        hash = (53 * hash) + getWeakDependencyList().hashCode();
-      }
-      if (getMessageTypeCount() > 0) {
-        hash = (37 * hash) + MESSAGE_TYPE_FIELD_NUMBER;
-        hash = (53 * hash) + getMessageTypeList().hashCode();
-      }
-      if (getEnumTypeCount() > 0) {
-        hash = (37 * hash) + ENUM_TYPE_FIELD_NUMBER;
-        hash = (53 * hash) + getEnumTypeList().hashCode();
-      }
-      if (getServiceCount() > 0) {
-        hash = (37 * hash) + SERVICE_FIELD_NUMBER;
-        hash = (53 * hash) + getServiceList().hashCode();
-      }
-      if (getExtensionCount() > 0) {
-        hash = (37 * hash) + EXTENSION_FIELD_NUMBER;
-        hash = (53 * hash) + getExtensionList().hashCode();
-      }
-      if (hasOptions()) {
-        hash = (37 * hash) + OPTIONS_FIELD_NUMBER;
-        hash = (53 * hash) + getOptions().hashCode();
-      }
-      if (hasSourceCodeInfo()) {
-        hash = (37 * hash) + SOURCE_CODE_INFO_FIELD_NUMBER;
-        hash = (53 * hash) + getSourceCodeInfo().hashCode();
-      }
-      if (hasSyntax()) {
-        hash = (37 * hash) + SYNTAX_FIELD_NUMBER;
-        hash = (53 * hash) + getSyntax().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto parseDelimitedFrom

<TRUNCATED>

[03/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AccessControlProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AccessControlProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AccessControlProtos.java
deleted file mode 100644
index e83a7ac..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AccessControlProtos.java
+++ /dev/null
@@ -1,11171 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: AccessControl.proto
-
-package org.apache.hadoop.hbase.shaded.protobuf.generated;
-
-public final class AccessControlProtos {
-  private AccessControlProtos() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  public interface PermissionOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:hbase.pb.Permission)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>required .hbase.pb.Permission.Type type = 1;</code>
-     */
-    boolean hasType();
-    /**
-     * <code>required .hbase.pb.Permission.Type type = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type getType();
-
-    /**
-     * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-     */
-    boolean hasGlobalPermission();
-    /**
-     * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission getGlobalPermission();
-    /**
-     * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermissionOrBuilder getGlobalPermissionOrBuilder();
-
-    /**
-     * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-     */
-    boolean hasNamespacePermission();
-    /**
-     * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission getNamespacePermission();
-    /**
-     * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermissionOrBuilder getNamespacePermissionOrBuilder();
-
-    /**
-     * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-     */
-    boolean hasTablePermission();
-    /**
-     * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission getTablePermission();
-    /**
-     * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermissionOrBuilder getTablePermissionOrBuilder();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.Permission}
-   */
-  public  static final class Permission extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:hbase.pb.Permission)
-      PermissionOrBuilder {
-    // Use Permission.newBuilder() to construct.
-    private Permission(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private Permission() {
-      type_ = 1;
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private Permission(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              int rawValue = input.readEnum();
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type value = org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type.valueOf(rawValue);
-              if (value == null) {
-                unknownFields.mergeVarintField(1, rawValue);
-              } else {
-                bitField0_ |= 0x00000001;
-                type_ = rawValue;
-              }
-              break;
-            }
-            case 18: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000002) == 0x00000002)) {
-                subBuilder = globalPermission_.toBuilder();
-              }
-              globalPermission_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(globalPermission_);
-                globalPermission_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000002;
-              break;
-            }
-            case 26: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000004) == 0x00000004)) {
-                subBuilder = namespacePermission_.toBuilder();
-              }
-              namespacePermission_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(namespacePermission_);
-                namespacePermission_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000004;
-              break;
-            }
-            case 34: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000008) == 0x00000008)) {
-                subBuilder = tablePermission_.toBuilder();
-              }
-              tablePermission_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(tablePermission_);
-                tablePermission_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000008;
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_Permission_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_Permission_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Builder.class);
-    }
-
-    /**
-     * Protobuf enum {@code hbase.pb.Permission.Action}
-     */
-    public enum Action
-        implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
-      /**
-       * <code>READ = 0;</code>
-       */
-      READ(0),
-      /**
-       * <code>WRITE = 1;</code>
-       */
-      WRITE(1),
-      /**
-       * <code>EXEC = 2;</code>
-       */
-      EXEC(2),
-      /**
-       * <code>CREATE = 3;</code>
-       */
-      CREATE(3),
-      /**
-       * <code>ADMIN = 4;</code>
-       */
-      ADMIN(4),
-      ;
-
-      /**
-       * <code>READ = 0;</code>
-       */
-      public static final int READ_VALUE = 0;
-      /**
-       * <code>WRITE = 1;</code>
-       */
-      public static final int WRITE_VALUE = 1;
-      /**
-       * <code>EXEC = 2;</code>
-       */
-      public static final int EXEC_VALUE = 2;
-      /**
-       * <code>CREATE = 3;</code>
-       */
-      public static final int CREATE_VALUE = 3;
-      /**
-       * <code>ADMIN = 4;</code>
-       */
-      public static final int ADMIN_VALUE = 4;
-
-
-      public final int getNumber() {
-        return value;
-      }
-
-      /**
-       * @deprecated Use {@link #forNumber(int)} instead.
-       */
-      @java.lang.Deprecated
-      public static Action valueOf(int value) {
-        return forNumber(value);
-      }
-
-      public static Action forNumber(int value) {
-        switch (value) {
-          case 0: return READ;
-          case 1: return WRITE;
-          case 2: return EXEC;
-          case 3: return CREATE;
-          case 4: return ADMIN;
-          default: return null;
-        }
-      }
-
-      public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Action>
-          internalGetValueMap() {
-        return internalValueMap;
-      }
-      private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
-          Action> internalValueMap =
-            new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Action>() {
-              public Action findValueByNumber(int number) {
-                return Action.forNumber(number);
-              }
-            };
-
-      public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
-          getValueDescriptor() {
-        return getDescriptor().getValues().get(ordinal());
-      }
-      public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-          getDescriptorForType() {
-        return getDescriptor();
-      }
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.getDescriptor().getEnumTypes().get(0);
-      }
-
-      private static final Action[] VALUES = values();
-
-      public static Action valueOf(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-        if (desc.getType() != getDescriptor()) {
-          throw new java.lang.IllegalArgumentException(
-            "EnumValueDescriptor is not for this type.");
-        }
-        return VALUES[desc.getIndex()];
-      }
-
-      private final int value;
-
-      private Action(int value) {
-        this.value = value;
-      }
-
-      // @@protoc_insertion_point(enum_scope:hbase.pb.Permission.Action)
-    }
-
-    /**
-     * Protobuf enum {@code hbase.pb.Permission.Type}
-     */
-    public enum Type
-        implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
-      /**
-       * <code>Global = 1;</code>
-       */
-      Global(1),
-      /**
-       * <code>Namespace = 2;</code>
-       */
-      Namespace(2),
-      /**
-       * <code>Table = 3;</code>
-       */
-      Table(3),
-      ;
-
-      /**
-       * <code>Global = 1;</code>
-       */
-      public static final int Global_VALUE = 1;
-      /**
-       * <code>Namespace = 2;</code>
-       */
-      public static final int Namespace_VALUE = 2;
-      /**
-       * <code>Table = 3;</code>
-       */
-      public static final int Table_VALUE = 3;
-
-
-      public final int getNumber() {
-        return value;
-      }
-
-      /**
-       * @deprecated Use {@link #forNumber(int)} instead.
-       */
-      @java.lang.Deprecated
-      public static Type valueOf(int value) {
-        return forNumber(value);
-      }
-
-      public static Type forNumber(int value) {
-        switch (value) {
-          case 1: return Global;
-          case 2: return Namespace;
-          case 3: return Table;
-          default: return null;
-        }
-      }
-
-      public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Type>
-          internalGetValueMap() {
-        return internalValueMap;
-      }
-      private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
-          Type> internalValueMap =
-            new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Type>() {
-              public Type findValueByNumber(int number) {
-                return Type.forNumber(number);
-              }
-            };
-
-      public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
-          getValueDescriptor() {
-        return getDescriptor().getValues().get(ordinal());
-      }
-      public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-          getDescriptorForType() {
-        return getDescriptor();
-      }
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.getDescriptor().getEnumTypes().get(1);
-      }
-
-      private static final Type[] VALUES = values();
-
-      public static Type valueOf(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-        if (desc.getType() != getDescriptor()) {
-          throw new java.lang.IllegalArgumentException(
-            "EnumValueDescriptor is not for this type.");
-        }
-        return VALUES[desc.getIndex()];
-      }
-
-      private final int value;
-
-      private Type(int value) {
-        this.value = value;
-      }
-
-      // @@protoc_insertion_point(enum_scope:hbase.pb.Permission.Type)
-    }
-
-    private int bitField0_;
-    public static final int TYPE_FIELD_NUMBER = 1;
-    private int type_;
-    /**
-     * <code>required .hbase.pb.Permission.Type type = 1;</code>
-     */
-    public boolean hasType() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required .hbase.pb.Permission.Type type = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type getType() {
-      org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type result = org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type.valueOf(type_);
-      return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type.Global : result;
-    }
-
-    public static final int GLOBAL_PERMISSION_FIELD_NUMBER = 2;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission globalPermission_;
-    /**
-     * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-     */
-    public boolean hasGlobalPermission() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission getGlobalPermission() {
-      return globalPermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.getDefaultInstance() : globalPermission_;
-    }
-    /**
-     * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermissionOrBuilder getGlobalPermissionOrBuilder() {
-      return globalPermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.getDefaultInstance() : globalPermission_;
-    }
-
-    public static final int NAMESPACE_PERMISSION_FIELD_NUMBER = 3;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission namespacePermission_;
-    /**
-     * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-     */
-    public boolean hasNamespacePermission() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission getNamespacePermission() {
-      return namespacePermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.getDefaultInstance() : namespacePermission_;
-    }
-    /**
-     * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermissionOrBuilder getNamespacePermissionOrBuilder() {
-      return namespacePermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.getDefaultInstance() : namespacePermission_;
-    }
-
-    public static final int TABLE_PERMISSION_FIELD_NUMBER = 4;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission tablePermission_;
-    /**
-     * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-     */
-    public boolean hasTablePermission() {
-      return ((bitField0_ & 0x00000008) == 0x00000008);
-    }
-    /**
-     * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission getTablePermission() {
-      return tablePermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.getDefaultInstance() : tablePermission_;
-    }
-    /**
-     * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermissionOrBuilder getTablePermissionOrBuilder() {
-      return tablePermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.getDefaultInstance() : tablePermission_;
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (!hasType()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (hasTablePermission()) {
-        if (!getTablePermission().isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeEnum(1, type_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeMessage(2, getGlobalPermission());
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeMessage(3, getNamespacePermission());
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        output.writeMessage(4, getTablePermission());
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeEnumSize(1, type_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(2, getGlobalPermission());
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(3, getNamespacePermission());
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(4, getTablePermission());
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission other = (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission) obj;
-
-      boolean result = true;
-      result = result && (hasType() == other.hasType());
-      if (hasType()) {
-        result = result && type_ == other.type_;
-      }
-      result = result && (hasGlobalPermission() == other.hasGlobalPermission());
-      if (hasGlobalPermission()) {
-        result = result && getGlobalPermission()
-            .equals(other.getGlobalPermission());
-      }
-      result = result && (hasNamespacePermission() == other.hasNamespacePermission());
-      if (hasNamespacePermission()) {
-        result = result && getNamespacePermission()
-            .equals(other.getNamespacePermission());
-      }
-      result = result && (hasTablePermission() == other.hasTablePermission());
-      if (hasTablePermission()) {
-        result = result && getTablePermission()
-            .equals(other.getTablePermission());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasType()) {
-        hash = (37 * hash) + TYPE_FIELD_NUMBER;
-        hash = (53 * hash) + type_;
-      }
-      if (hasGlobalPermission()) {
-        hash = (37 * hash) + GLOBAL_PERMISSION_FIELD_NUMBER;
-        hash = (53 * hash) + getGlobalPermission().hashCode();
-      }
-      if (hasNamespacePermission()) {
-        hash = (37 * hash) + NAMESPACE_PERMISSION_FIELD_NUMBER;
-        hash = (53 * hash) + getNamespacePermission().hashCode();
-      }
-      if (hasTablePermission()) {
-        hash = (37 * hash) + TABLE_PERMISSION_FIELD_NUMBER;
-        hash = (53 * hash) + getTablePermission().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.Permission}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:hbase.pb.Permission)
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.PermissionOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_Permission_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_Permission_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-          getGlobalPermissionFieldBuilder();
-          getNamespacePermissionFieldBuilder();
-          getTablePermissionFieldBuilder();
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        type_ = 1;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        if (globalPermissionBuilder_ == null) {
-          globalPermission_ = null;
-        } else {
-          globalPermissionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000002);
-        if (namespacePermissionBuilder_ == null) {
-          namespacePermission_ = null;
-        } else {
-          namespacePermissionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000004);
-        if (tablePermissionBuilder_ == null) {
-          tablePermission_ = null;
-        } else {
-          tablePermissionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000008);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_Permission_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission build() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission buildPartial() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission result = new org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.type_ = type_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        if (globalPermissionBuilder_ == null) {
-          result.globalPermission_ = globalPermission_;
-        } else {
-          result.globalPermission_ = globalPermissionBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        if (namespacePermissionBuilder_ == null) {
-          result.namespacePermission_ = namespacePermission_;
-        } else {
-          result.namespacePermission_ = namespacePermissionBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
-          to_bitField0_ |= 0x00000008;
-        }
-        if (tablePermissionBuilder_ == null) {
-          result.tablePermission_ = tablePermission_;
-        } else {
-          result.tablePermission_ = tablePermissionBuilder_.build();
-        }
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission other) {
-        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.getDefaultInstance()) return this;
-        if (other.hasType()) {
-          setType(other.getType());
-        }
-        if (other.hasGlobalPermission()) {
-          mergeGlobalPermission(other.getGlobalPermission());
-        }
-        if (other.hasNamespacePermission()) {
-          mergeNamespacePermission(other.getNamespacePermission());
-        }
-        if (other.hasTablePermission()) {
-          mergeTablePermission(other.getTablePermission());
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasType()) {
-          return false;
-        }
-        if (hasTablePermission()) {
-          if (!getTablePermission().isInitialized()) {
-            return false;
-          }
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private int type_ = 1;
-      /**
-       * <code>required .hbase.pb.Permission.Type type = 1;</code>
-       */
-      public boolean hasType() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required .hbase.pb.Permission.Type type = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type getType() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type result = org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type.valueOf(type_);
-        return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type.Global : result;
-      }
-      /**
-       * <code>required .hbase.pb.Permission.Type type = 1;</code>
-       */
-      public Builder setType(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Type value) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        bitField0_ |= 0x00000001;
-        type_ = value.getNumber();
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Permission.Type type = 1;</code>
-       */
-      public Builder clearType() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        type_ = 1;
-        onChanged();
-        return this;
-      }
-
-      private org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission globalPermission_ = null;
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermissionOrBuilder> globalPermissionBuilder_;
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      public boolean hasGlobalPermission() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission getGlobalPermission() {
-        if (globalPermissionBuilder_ == null) {
-          return globalPermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.getDefaultInstance() : globalPermission_;
-        } else {
-          return globalPermissionBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      public Builder setGlobalPermission(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission value) {
-        if (globalPermissionBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          globalPermission_ = value;
-          onChanged();
-        } else {
-          globalPermissionBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000002;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      public Builder setGlobalPermission(
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.Builder builderForValue) {
-        if (globalPermissionBuilder_ == null) {
-          globalPermission_ = builderForValue.build();
-          onChanged();
-        } else {
-          globalPermissionBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000002;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      public Builder mergeGlobalPermission(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission value) {
-        if (globalPermissionBuilder_ == null) {
-          if (((bitField0_ & 0x00000002) == 0x00000002) &&
-              globalPermission_ != null &&
-              globalPermission_ != org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.getDefaultInstance()) {
-            globalPermission_ =
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.newBuilder(globalPermission_).mergeFrom(value).buildPartial();
-          } else {
-            globalPermission_ = value;
-          }
-          onChanged();
-        } else {
-          globalPermissionBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000002;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      public Builder clearGlobalPermission() {
-        if (globalPermissionBuilder_ == null) {
-          globalPermission_ = null;
-          onChanged();
-        } else {
-          globalPermissionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.Builder getGlobalPermissionBuilder() {
-        bitField0_ |= 0x00000002;
-        onChanged();
-        return getGlobalPermissionFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermissionOrBuilder getGlobalPermissionOrBuilder() {
-        if (globalPermissionBuilder_ != null) {
-          return globalPermissionBuilder_.getMessageOrBuilder();
-        } else {
-          return globalPermission_ == null ?
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.getDefaultInstance() : globalPermission_;
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.GlobalPermission global_permission = 2;</code>
-       */
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermissionOrBuilder> 
-          getGlobalPermissionFieldBuilder() {
-        if (globalPermissionBuilder_ == null) {
-          globalPermissionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GlobalPermissionOrBuilder>(
-                  getGlobalPermission(),
-                  getParentForChildren(),
-                  isClean());
-          globalPermission_ = null;
-        }
-        return globalPermissionBuilder_;
-      }
-
-      private org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission namespacePermission_ = null;
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermissionOrBuilder> namespacePermissionBuilder_;
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      public boolean hasNamespacePermission() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
-      }
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission getNamespacePermission() {
-        if (namespacePermissionBuilder_ == null) {
-          return namespacePermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.getDefaultInstance() : namespacePermission_;
-        } else {
-          return namespacePermissionBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      public Builder setNamespacePermission(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission value) {
-        if (namespacePermissionBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          namespacePermission_ = value;
-          onChanged();
-        } else {
-          namespacePermissionBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000004;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      public Builder setNamespacePermission(
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.Builder builderForValue) {
-        if (namespacePermissionBuilder_ == null) {
-          namespacePermission_ = builderForValue.build();
-          onChanged();
-        } else {
-          namespacePermissionBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000004;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      public Builder mergeNamespacePermission(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission value) {
-        if (namespacePermissionBuilder_ == null) {
-          if (((bitField0_ & 0x00000004) == 0x00000004) &&
-              namespacePermission_ != null &&
-              namespacePermission_ != org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.getDefaultInstance()) {
-            namespacePermission_ =
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.newBuilder(namespacePermission_).mergeFrom(value).buildPartial();
-          } else {
-            namespacePermission_ = value;
-          }
-          onChanged();
-        } else {
-          namespacePermissionBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000004;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      public Builder clearNamespacePermission() {
-        if (namespacePermissionBuilder_ == null) {
-          namespacePermission_ = null;
-          onChanged();
-        } else {
-          namespacePermissionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000004);
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.Builder getNamespacePermissionBuilder() {
-        bitField0_ |= 0x00000004;
-        onChanged();
-        return getNamespacePermissionFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermissionOrBuilder getNamespacePermissionOrBuilder() {
-        if (namespacePermissionBuilder_ != null) {
-          return namespacePermissionBuilder_.getMessageOrBuilder();
-        } else {
-          return namespacePermission_ == null ?
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.getDefaultInstance() : namespacePermission_;
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.NamespacePermission namespace_permission = 3;</code>
-       */
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermissionOrBuilder> 
-          getNamespacePermissionFieldBuilder() {
-        if (namespacePermissionBuilder_ == null) {
-          namespacePermissionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.NamespacePermissionOrBuilder>(
-                  getNamespacePermission(),
-                  getParentForChildren(),
-                  isClean());
-          namespacePermission_ = null;
-        }
-        return namespacePermissionBuilder_;
-      }
-
-      private org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission tablePermission_ = null;
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermissionOrBuilder> tablePermissionBuilder_;
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      public boolean hasTablePermission() {
-        return ((bitField0_ & 0x00000008) == 0x00000008);
-      }
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission getTablePermission() {
-        if (tablePermissionBuilder_ == null) {
-          return tablePermission_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.getDefaultInstance() : tablePermission_;
-        } else {
-          return tablePermissionBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      public Builder setTablePermission(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission value) {
-        if (tablePermissionBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          tablePermission_ = value;
-          onChanged();
-        } else {
-          tablePermissionBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000008;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      public Builder setTablePermission(
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.Builder builderForValue) {
-        if (tablePermissionBuilder_ == null) {
-          tablePermission_ = builderForValue.build();
-          onChanged();
-        } else {
-          tablePermissionBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000008;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      public Builder mergeTablePermission(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission value) {
-        if (tablePermissionBuilder_ == null) {
-          if (((bitField0_ & 0x00000008) == 0x00000008) &&
-              tablePermission_ != null &&
-              tablePermission_ != org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.getDefaultInstance()) {
-            tablePermission_ =
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.newBuilder(tablePermission_).mergeFrom(value).buildPartial();
-          } else {
-            tablePermission_ = value;
-          }
-          onChanged();
-        } else {
-          tablePermissionBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000008;
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      public Builder clearTablePermission() {
-        if (tablePermissionBuilder_ == null) {
-          tablePermission_ = null;
-          onChanged();
-        } else {
-          tablePermissionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000008);
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.Builder getTablePermissionBuilder() {
-        bitField0_ |= 0x00000008;
-        onChanged();
-        return getTablePermissionFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermissionOrBuilder getTablePermissionOrBuilder() {
-        if (tablePermissionBuilder_ != null) {
-          return tablePermissionBuilder_.getMessageOrBuilder();
-        } else {
-          return tablePermission_ == null ?
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.getDefaultInstance() : tablePermission_;
-        }
-      }
-      /**
-       * <code>optional .hbase.pb.TablePermission table_permission = 4;</code>
-       */
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermissionOrBuilder> 
-          getTablePermissionFieldBuilder() {
-        if (tablePermissionBuilder_ == null) {
-          tablePermissionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermissionOrBuilder>(
-                  getTablePermission(),
-                  getParentForChildren(),
-                  isClean());
-          tablePermission_ = null;
-        }
-        return tablePermissionBuilder_;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.Permission)
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.Permission)
-    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Permission>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Permission>() {
-      public Permission parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new Permission(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Permission> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Permission> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface TablePermissionOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:hbase.pb.TablePermission)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    boolean hasTableName();
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName();
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();
-
-    /**
-     * <code>optional bytes family = 2;</code>
-     */
-    boolean hasFamily();
-    /**
-     * <code>optional bytes family = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily();
-
-    /**
-     * <code>optional bytes qualifier = 3;</code>
-     */
-    boolean hasQualifier();
-    /**
-     * <code>optional bytes qualifier = 3;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier();
-
-    /**
-     * <code>repeated .hbase.pb.Permission.Action action = 4;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action> getActionList();
-    /**
-     * <code>repeated .hbase.pb.Permission.Action action = 4;</code>
-     */
-    int getActionCount();
-    /**
-     * <code>repeated .hbase.pb.Permission.Action action = 4;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action getAction(int index);
-  }
-  /**
-   * Protobuf type {@code hbase.pb.TablePermission}
-   */
-  public  static final class TablePermission extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:hbase.pb.TablePermission)
-      TablePermissionOrBuilder {
-    // Use TablePermission.newBuilder() to construct.
-    private TablePermission(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private TablePermission() {
-      family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-      qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-      action_ = java.util.Collections.emptyList();
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private TablePermission(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = tableName_.toBuilder();
-              }
-              tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(tableName_);
-                tableName_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000001;
-              break;
-            }
-            case 18: {
-              bitField0_ |= 0x00000002;
-              family_ = input.readBytes();
-              break;
-            }
-            case 26: {
-              bitField0_ |= 0x00000004;
-              qualifier_ = input.readBytes();
-              break;
-            }
-            case 32: {
-              int rawValue = input.readEnum();
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action value = org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action.valueOf(rawValue);
-              if (value == null) {
-                unknownFields.mergeVarintField(4, rawValue);
-              } else {
-                if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-                  action_ = new java.util.ArrayList<java.lang.Integer>();
-                  mutable_bitField0_ |= 0x00000008;
-                }
-                action_.add(rawValue);
-              }
-              break;
-            }
-            case 34: {
-              int length = input.readRawVarint32();
-              int oldLimit = input.pushLimit(length);
-              while(input.getBytesUntilLimit() > 0) {
-                int rawValue = input.readEnum();
-                org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action value = org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action.valueOf(rawValue);
-                if (value == null) {
-                  unknownFields.mergeVarintField(4, rawValue);
-                } else {
-                  if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-                    action_ = new java.util.ArrayList<java.lang.Integer>();
-                    mutable_bitField0_ |= 0x00000008;
-                  }
-                  action_.add(rawValue);
-                }
-              }
-              input.popLimit(oldLimit);
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
-          action_ = java.util.Collections.unmodifiableList(action_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_TablePermission_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_TablePermission_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int TABLE_NAME_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    public boolean hasTableName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
-      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
-    }
-    /**
-     * <code>optional .hbase.pb.TableName table_name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
-      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
-    }
-
-    public static final int FAMILY_FIELD_NUMBER = 2;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString family_;
-    /**
-     * <code>optional bytes family = 2;</code>
-     */
-    public boolean hasFamily() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional bytes family = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily() {
-      return family_;
-    }
-
-    public static final int QUALIFIER_FIELD_NUMBER = 3;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString qualifier_;
-    /**
-     * <code>optional bytes qualifier = 3;</code>
-     */
-    public boolean hasQualifier() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>optional bytes qualifier = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getQualifier() {
-      return qualifier_;
-    }
-
-    public static final int ACTION_FIELD_NUMBER = 4;
-    private java.util.List<java.lang.Integer> action_;
-    private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.ListAdapter.Converter<
-        java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action> action_converter_ =
-            new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.ListAdapter.Converter<
-                java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action>() {
-              public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action convert(java.lang.Integer from) {
-                org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action result = org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action.valueOf(from);
-                return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action.READ : result;
-              }
-            };
-    /**
-     * <code>repeated .hbase.pb.Permission.Action action = 4;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action> getActionList() {
-      return new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.ListAdapter<
-          java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action>(action_, action_converter_);
-    }
-    /**
-     * <code>repeated .hbase.pb.Permission.Action action = 4;</code>
-     */
-    public int getActionCount() {
-      return action_.size();
-    }
-    /**
-     * <code>repeated .hbase.pb.Permission.Action action = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action getAction(int index) {
-      return action_converter_.convert(action_.get(index));
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (hasTableName()) {
-        if (!getTableName().isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, getTableName());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeBytes(2, family_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeBytes(3, qualifier_);
-      }
-      for (int i = 0; i < action_.size(); i++) {
-        output.writeEnum(4, action_.get(i));
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, getTableName());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeBytesSize(2, family_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeBytesSize(3, qualifier_);
-      }
-      {
-        int dataSize = 0;
-        for (int i = 0; i < action_.size(); i++) {
-          dataSize += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-            .computeEnumSizeNoTag(action_.get(i));
-        }
-        size += dataSize;
-        size += 1 * action_.size();
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission other = (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission) obj;
-
-      boolean result = true;
-      result = result && (hasTableName() == other.hasTableName());
-      if (hasTableName()) {
-        result = result && getTableName()
-            .equals(other.getTableName());
-      }
-      result = result && (hasFamily() == other.hasFamily());
-      if (hasFamily()) {
-        result = result && getFamily()
-            .equals(other.getFamily());
-      }
-      result = result && (hasQualifier() == other.hasQualifier());
-      if (hasQualifier()) {
-        result = result && getQualifier()
-            .equals(other.getQualifier());
-      }
-      result = result && action_.equals(other.action_);
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasTableName()) {
-        hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
-        hash = (53 * hash) + getTableName().hashCode();
-      }
-      if (hasFamily()) {
-        hash = (37 * hash) + FAMILY_FIELD_NUMBER;
-        hash = (53 * hash) + getFamily().hashCode();
-      }
-      if (hasQualifier()) {
-        hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
-        hash = (53 * hash) + getQualifier().hashCode();
-      }
-      if (getActionCount() > 0) {
-        hash = (37 * hash) + ACTION_FIELD_NUMBER;
-        hash = (53 * hash) + action_.hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.TablePermission}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:hbase.pb.TablePermission)
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermissionOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_TablePermission_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_TablePermission_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-          getTableNameFieldBuilder();
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        if (tableNameBuilder_ == null) {
-          tableName_ = null;
-        } else {
-          tableNameBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        family_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        qualifier_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000004);
-        action_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000008);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.internal_static_hbase_pb_TablePermission_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission build() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission buildPartial() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission result = new org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        if (tableNameBuilder_ == null) {
-          result.tableName_ = tableName_;
-        } else {
-          result.tableName_ = tableNameBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.family_ = family_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        result.qualifier_ = qualifier_;
-        if (((bitField0_ & 0x00000008) == 0x00000008)) {
-          action_ = java.util.Collections.unmodifiableList(action_);
-          bitField0_ = (bitField0_ & ~0x00000008);
-        }
-        result.action_ = action_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission other) {
-        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission.getDefaultInstance()) return this;
-        if (other.hasTableName()) {
-          mergeTableName(other.getTableName());
-        }
-        if (other.hasFamily()) {
-          setFamily(other.getFamily());
-        }
-        if (other.hasQualifier()) {
-          setQualifier(other.getQualifier());
-        }
-        if (!other.action_.isEmpty()) {
-          if (action_.isEmpty()) {
-            action_ = other.action_;
-            bitField0_ = (bitField0_ & ~0x00000008);
-          } else {
-            ensureActionIsMutable();
-            action_.addAll(other.action_);
-          }
-          onChanged();
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (hasTableName()) {
-          if (!getTableName().isInitialized()) {
-            return false;
-          }
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.TablePermission parsedMessage

<TRUNCATED>

[50/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/BulkDeleteProtos.java
----------------------------------------------------------------------
diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/BulkDeleteProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/BulkDeleteProtos.java
deleted file mode 100644
index 373e036..0000000
--- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/BulkDeleteProtos.java
+++ /dev/null
@@ -1,1792 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: BulkDelete.proto
-
-package org.apache.hadoop.hbase.coprocessor.example.generated;
-
-public final class BulkDeleteProtos {
-  private BulkDeleteProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface BulkDeleteRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required .hbase.pb.Scan scan = 1;
-    /**
-     * <code>required .hbase.pb.Scan scan = 1;</code>
-     */
-    boolean hasScan();
-    /**
-     * <code>required .hbase.pb.Scan scan = 1;</code>
-     */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan();
-    /**
-     * <code>required .hbase.pb.Scan scan = 1;</code>
-     */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder();
-
-    // required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;
-    /**
-     * <code>required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;</code>
-     */
-    boolean hasDeleteType();
-    /**
-     * <code>required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;</code>
-     */
-    org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType getDeleteType();
-
-    // optional uint64 timestamp = 3;
-    /**
-     * <code>optional uint64 timestamp = 3;</code>
-     */
-    boolean hasTimestamp();
-    /**
-     * <code>optional uint64 timestamp = 3;</code>
-     */
-    long getTimestamp();
-
-    // required uint32 rowBatchSize = 4;
-    /**
-     * <code>required uint32 rowBatchSize = 4;</code>
-     */
-    boolean hasRowBatchSize();
-    /**
-     * <code>required uint32 rowBatchSize = 4;</code>
-     */
-    int getRowBatchSize();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.BulkDeleteRequest}
-   */
-  public static final class BulkDeleteRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements BulkDeleteRequestOrBuilder {
-    // Use BulkDeleteRequest.newBuilder() to construct.
-    private BulkDeleteRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private BulkDeleteRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final BulkDeleteRequest defaultInstance;
-    public static BulkDeleteRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public BulkDeleteRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private BulkDeleteRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = scan_.toBuilder();
-              }
-              scan_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(scan_);
-                scan_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000001;
-              break;
-            }
-            case 16: {
-              int rawValue = input.readEnum();
-              org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType value = org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType.valueOf(rawValue);
-              if (value == null) {
-                unknownFields.mergeVarintField(2, rawValue);
-              } else {
-                bitField0_ |= 0x00000002;
-                deleteType_ = value;
-              }
-              break;
-            }
-            case 24: {
-              bitField0_ |= 0x00000004;
-              timestamp_ = input.readUInt64();
-              break;
-            }
-            case 32: {
-              bitField0_ |= 0x00000008;
-              rowBatchSize_ = input.readUInt32();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.class, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<BulkDeleteRequest> PARSER =
-        new com.google.protobuf.AbstractParser<BulkDeleteRequest>() {
-      public BulkDeleteRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new BulkDeleteRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<BulkDeleteRequest> getParserForType() {
-      return PARSER;
-    }
-
-    /**
-     * Protobuf enum {@code hbase.pb.BulkDeleteRequest.DeleteType}
-     */
-    public enum DeleteType
-        implements com.google.protobuf.ProtocolMessageEnum {
-      /**
-       * <code>ROW = 0;</code>
-       */
-      ROW(0, 0),
-      /**
-       * <code>FAMILY = 1;</code>
-       */
-      FAMILY(1, 1),
-      /**
-       * <code>COLUMN = 2;</code>
-       */
-      COLUMN(2, 2),
-      /**
-       * <code>VERSION = 3;</code>
-       */
-      VERSION(3, 3),
-      ;
-
-      /**
-       * <code>ROW = 0;</code>
-       */
-      public static final int ROW_VALUE = 0;
-      /**
-       * <code>FAMILY = 1;</code>
-       */
-      public static final int FAMILY_VALUE = 1;
-      /**
-       * <code>COLUMN = 2;</code>
-       */
-      public static final int COLUMN_VALUE = 2;
-      /**
-       * <code>VERSION = 3;</code>
-       */
-      public static final int VERSION_VALUE = 3;
-
-
-      public final int getNumber() { return value; }
-
-      public static DeleteType valueOf(int value) {
-        switch (value) {
-          case 0: return ROW;
-          case 1: return FAMILY;
-          case 2: return COLUMN;
-          case 3: return VERSION;
-          default: return null;
-        }
-      }
-
-      public static com.google.protobuf.Internal.EnumLiteMap<DeleteType>
-          internalGetValueMap() {
-        return internalValueMap;
-      }
-      private static com.google.protobuf.Internal.EnumLiteMap<DeleteType>
-          internalValueMap =
-            new com.google.protobuf.Internal.EnumLiteMap<DeleteType>() {
-              public DeleteType findValueByNumber(int number) {
-                return DeleteType.valueOf(number);
-              }
-            };
-
-      public final com.google.protobuf.Descriptors.EnumValueDescriptor
-          getValueDescriptor() {
-        return getDescriptor().getValues().get(index);
-      }
-      public final com.google.protobuf.Descriptors.EnumDescriptor
-          getDescriptorForType() {
-        return getDescriptor();
-      }
-      public static final com.google.protobuf.Descriptors.EnumDescriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.getDescriptor().getEnumTypes().get(0);
-      }
-
-      private static final DeleteType[] VALUES = values();
-
-      public static DeleteType valueOf(
-          com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-        if (desc.getType() != getDescriptor()) {
-          throw new java.lang.IllegalArgumentException(
-            "EnumValueDescriptor is not for this type.");
-        }
-        return VALUES[desc.getIndex()];
-      }
-
-      private final int index;
-      private final int value;
-
-      private DeleteType(int index, int value) {
-        this.index = index;
-        this.value = value;
-      }
-
-      // @@protoc_insertion_point(enum_scope:hbase.pb.BulkDeleteRequest.DeleteType)
-    }
-
-    private int bitField0_;
-    // required .hbase.pb.Scan scan = 1;
-    public static final int SCAN_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan scan_;
-    /**
-     * <code>required .hbase.pb.Scan scan = 1;</code>
-     */
-    public boolean hasScan() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required .hbase.pb.Scan scan = 1;</code>
-     */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan() {
-      return scan_;
-    }
-    /**
-     * <code>required .hbase.pb.Scan scan = 1;</code>
-     */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder() {
-      return scan_;
-    }
-
-    // required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;
-    public static final int DELETETYPE_FIELD_NUMBER = 2;
-    private org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType deleteType_;
-    /**
-     * <code>required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;</code>
-     */
-    public boolean hasDeleteType() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;</code>
-     */
-    public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType getDeleteType() {
-      return deleteType_;
-    }
-
-    // optional uint64 timestamp = 3;
-    public static final int TIMESTAMP_FIELD_NUMBER = 3;
-    private long timestamp_;
-    /**
-     * <code>optional uint64 timestamp = 3;</code>
-     */
-    public boolean hasTimestamp() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>optional uint64 timestamp = 3;</code>
-     */
-    public long getTimestamp() {
-      return timestamp_;
-    }
-
-    // required uint32 rowBatchSize = 4;
-    public static final int ROWBATCHSIZE_FIELD_NUMBER = 4;
-    private int rowBatchSize_;
-    /**
-     * <code>required uint32 rowBatchSize = 4;</code>
-     */
-    public boolean hasRowBatchSize() {
-      return ((bitField0_ & 0x00000008) == 0x00000008);
-    }
-    /**
-     * <code>required uint32 rowBatchSize = 4;</code>
-     */
-    public int getRowBatchSize() {
-      return rowBatchSize_;
-    }
-
-    private void initFields() {
-      scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance();
-      deleteType_ = org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType.ROW;
-      timestamp_ = 0L;
-      rowBatchSize_ = 0;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasScan()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!hasDeleteType()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!hasRowBatchSize()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!getScan().isInitialized()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, scan_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeEnum(2, deleteType_.getNumber());
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeUInt64(3, timestamp_);
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        output.writeUInt32(4, rowBatchSize_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, scan_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeEnumSize(2, deleteType_.getNumber());
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeUInt64Size(3, timestamp_);
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeUInt32Size(4, rowBatchSize_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest other = (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest) obj;
-
-      boolean result = true;
-      result = result && (hasScan() == other.hasScan());
-      if (hasScan()) {
-        result = result && getScan()
-            .equals(other.getScan());
-      }
-      result = result && (hasDeleteType() == other.hasDeleteType());
-      if (hasDeleteType()) {
-        result = result &&
-            (getDeleteType() == other.getDeleteType());
-      }
-      result = result && (hasTimestamp() == other.hasTimestamp());
-      if (hasTimestamp()) {
-        result = result && (getTimestamp()
-            == other.getTimestamp());
-      }
-      result = result && (hasRowBatchSize() == other.hasRowBatchSize());
-      if (hasRowBatchSize()) {
-        result = result && (getRowBatchSize()
-            == other.getRowBatchSize());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasScan()) {
-        hash = (37 * hash) + SCAN_FIELD_NUMBER;
-        hash = (53 * hash) + getScan().hashCode();
-      }
-      if (hasDeleteType()) {
-        hash = (37 * hash) + DELETETYPE_FIELD_NUMBER;
-        hash = (53 * hash) + hashEnum(getDeleteType());
-      }
-      if (hasTimestamp()) {
-        hash = (37 * hash) + TIMESTAMP_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getTimestamp());
-      }
-      if (hasRowBatchSize()) {
-        hash = (37 * hash) + ROWBATCHSIZE_FIELD_NUMBER;
-        hash = (53 * hash) + getRowBatchSize();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.BulkDeleteRequest}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.class, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-          getScanFieldBuilder();
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        if (scanBuilder_ == null) {
-          scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance();
-        } else {
-          scanBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        deleteType_ = org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType.ROW;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        timestamp_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000004);
-        rowBatchSize_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000008);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest build() {
-        org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest result = new org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        if (scanBuilder_ == null) {
-          result.scan_ = scan_;
-        } else {
-          result.scan_ = scanBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.deleteType_ = deleteType_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        result.timestamp_ = timestamp_;
-        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
-          to_bitField0_ |= 0x00000008;
-        }
-        result.rowBatchSize_ = rowBatchSize_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.getDefaultInstance()) return this;
-        if (other.hasScan()) {
-          mergeScan(other.getScan());
-        }
-        if (other.hasDeleteType()) {
-          setDeleteType(other.getDeleteType());
-        }
-        if (other.hasTimestamp()) {
-          setTimestamp(other.getTimestamp());
-        }
-        if (other.hasRowBatchSize()) {
-          setRowBatchSize(other.getRowBatchSize());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasScan()) {
-          
-          return false;
-        }
-        if (!hasDeleteType()) {
-          
-          return false;
-        }
-        if (!hasRowBatchSize()) {
-          
-          return false;
-        }
-        if (!getScan().isInitialized()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required .hbase.pb.Scan scan = 1;
-      private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder> scanBuilder_;
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      public boolean hasScan() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan getScan() {
-        if (scanBuilder_ == null) {
-          return scan_;
-        } else {
-          return scanBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      public Builder setScan(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan value) {
-        if (scanBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          scan_ = value;
-          onChanged();
-        } else {
-          scanBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      public Builder setScan(
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder builderForValue) {
-        if (scanBuilder_ == null) {
-          scan_ = builderForValue.build();
-          onChanged();
-        } else {
-          scanBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      public Builder mergeScan(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan value) {
-        if (scanBuilder_ == null) {
-          if (((bitField0_ & 0x00000001) == 0x00000001) &&
-              scan_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance()) {
-            scan_ =
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.newBuilder(scan_).mergeFrom(value).buildPartial();
-          } else {
-            scan_ = value;
-          }
-          onChanged();
-        } else {
-          scanBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      public Builder clearScan() {
-        if (scanBuilder_ == null) {
-          scan_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.getDefaultInstance();
-          onChanged();
-        } else {
-          scanBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder getScanBuilder() {
-        bitField0_ |= 0x00000001;
-        onChanged();
-        return getScanFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder getScanOrBuilder() {
-        if (scanBuilder_ != null) {
-          return scanBuilder_.getMessageOrBuilder();
-        } else {
-          return scan_;
-        }
-      }
-      /**
-       * <code>required .hbase.pb.Scan scan = 1;</code>
-       */
-      private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder> 
-          getScanFieldBuilder() {
-        if (scanBuilder_ == null) {
-          scanBuilder_ = new com.google.protobuf.SingleFieldBuilder<
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Scan.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ScanOrBuilder>(
-                  scan_,
-                  getParentForChildren(),
-                  isClean());
-          scan_ = null;
-        }
-        return scanBuilder_;
-      }
-
-      // required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;
-      private org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType deleteType_ = org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType.ROW;
-      /**
-       * <code>required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;</code>
-       */
-      public boolean hasDeleteType() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;</code>
-       */
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType getDeleteType() {
-        return deleteType_;
-      }
-      /**
-       * <code>required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;</code>
-       */
-      public Builder setDeleteType(org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType value) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        bitField0_ |= 0x00000002;
-        deleteType_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.BulkDeleteRequest.DeleteType deleteType = 2;</code>
-       */
-      public Builder clearDeleteType() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        deleteType_ = org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.DeleteType.ROW;
-        onChanged();
-        return this;
-      }
-
-      // optional uint64 timestamp = 3;
-      private long timestamp_ ;
-      /**
-       * <code>optional uint64 timestamp = 3;</code>
-       */
-      public boolean hasTimestamp() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
-      }
-      /**
-       * <code>optional uint64 timestamp = 3;</code>
-       */
-      public long getTimestamp() {
-        return timestamp_;
-      }
-      /**
-       * <code>optional uint64 timestamp = 3;</code>
-       */
-      public Builder setTimestamp(long value) {
-        bitField0_ |= 0x00000004;
-        timestamp_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional uint64 timestamp = 3;</code>
-       */
-      public Builder clearTimestamp() {
-        bitField0_ = (bitField0_ & ~0x00000004);
-        timestamp_ = 0L;
-        onChanged();
-        return this;
-      }
-
-      // required uint32 rowBatchSize = 4;
-      private int rowBatchSize_ ;
-      /**
-       * <code>required uint32 rowBatchSize = 4;</code>
-       */
-      public boolean hasRowBatchSize() {
-        return ((bitField0_ & 0x00000008) == 0x00000008);
-      }
-      /**
-       * <code>required uint32 rowBatchSize = 4;</code>
-       */
-      public int getRowBatchSize() {
-        return rowBatchSize_;
-      }
-      /**
-       * <code>required uint32 rowBatchSize = 4;</code>
-       */
-      public Builder setRowBatchSize(int value) {
-        bitField0_ |= 0x00000008;
-        rowBatchSize_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required uint32 rowBatchSize = 4;</code>
-       */
-      public Builder clearRowBatchSize() {
-        bitField0_ = (bitField0_ & ~0x00000008);
-        rowBatchSize_ = 0;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.BulkDeleteRequest)
-    }
-
-    static {
-      defaultInstance = new BulkDeleteRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.BulkDeleteRequest)
-  }
-
-  public interface BulkDeleteResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required uint64 rowsDeleted = 1;
-    /**
-     * <code>required uint64 rowsDeleted = 1;</code>
-     */
-    boolean hasRowsDeleted();
-    /**
-     * <code>required uint64 rowsDeleted = 1;</code>
-     */
-    long getRowsDeleted();
-
-    // optional uint64 versionsDeleted = 2;
-    /**
-     * <code>optional uint64 versionsDeleted = 2;</code>
-     */
-    boolean hasVersionsDeleted();
-    /**
-     * <code>optional uint64 versionsDeleted = 2;</code>
-     */
-    long getVersionsDeleted();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.BulkDeleteResponse}
-   */
-  public static final class BulkDeleteResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements BulkDeleteResponseOrBuilder {
-    // Use BulkDeleteResponse.newBuilder() to construct.
-    private BulkDeleteResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private BulkDeleteResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final BulkDeleteResponse defaultInstance;
-    public static BulkDeleteResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public BulkDeleteResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private BulkDeleteResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              rowsDeleted_ = input.readUInt64();
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              versionsDeleted_ = input.readUInt64();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<BulkDeleteResponse> PARSER =
-        new com.google.protobuf.AbstractParser<BulkDeleteResponse>() {
-      public BulkDeleteResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new BulkDeleteResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<BulkDeleteResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required uint64 rowsDeleted = 1;
-    public static final int ROWSDELETED_FIELD_NUMBER = 1;
-    private long rowsDeleted_;
-    /**
-     * <code>required uint64 rowsDeleted = 1;</code>
-     */
-    public boolean hasRowsDeleted() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required uint64 rowsDeleted = 1;</code>
-     */
-    public long getRowsDeleted() {
-      return rowsDeleted_;
-    }
-
-    // optional uint64 versionsDeleted = 2;
-    public static final int VERSIONSDELETED_FIELD_NUMBER = 2;
-    private long versionsDeleted_;
-    /**
-     * <code>optional uint64 versionsDeleted = 2;</code>
-     */
-    public boolean hasVersionsDeleted() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional uint64 versionsDeleted = 2;</code>
-     */
-    public long getVersionsDeleted() {
-      return versionsDeleted_;
-    }
-
-    private void initFields() {
-      rowsDeleted_ = 0L;
-      versionsDeleted_ = 0L;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasRowsDeleted()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeUInt64(1, rowsDeleted_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeUInt64(2, versionsDeleted_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeUInt64Size(1, rowsDeleted_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeUInt64Size(2, versionsDeleted_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse other = (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse) obj;
-
-      boolean result = true;
-      result = result && (hasRowsDeleted() == other.hasRowsDeleted());
-      if (hasRowsDeleted()) {
-        result = result && (getRowsDeleted()
-            == other.getRowsDeleted());
-      }
-      result = result && (hasVersionsDeleted() == other.hasVersionsDeleted());
-      if (hasVersionsDeleted()) {
-        result = result && (getVersionsDeleted()
-            == other.getVersionsDeleted());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasRowsDeleted()) {
-        hash = (37 * hash) + ROWSDELETED_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getRowsDeleted());
-      }
-      if (hasVersionsDeleted()) {
-        hash = (37 * hash) + VERSIONSDELETED_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getVersionsDeleted());
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.BulkDeleteResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        rowsDeleted_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        versionsDeleted_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.internal_static_hbase_pb_BulkDeleteResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse build() {
-        org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse result = new org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.rowsDeleted_ = rowsDeleted_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.versionsDeleted_ = versionsDeleted_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance()) return this;
-        if (other.hasRowsDeleted()) {
-          setRowsDeleted(other.getRowsDeleted());
-        }
-        if (other.hasVersionsDeleted()) {
-          setVersionsDeleted(other.getVersionsDeleted());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasRowsDeleted()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required uint64 rowsDeleted = 1;
-      private long rowsDeleted_ ;
-      /**
-       * <code>required uint64 rowsDeleted = 1;</code>
-       */
-      public boolean hasRowsDeleted() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required uint64 rowsDeleted = 1;</code>
-       */
-      public long getRowsDeleted() {
-        return rowsDeleted_;
-      }
-      /**
-       * <code>required uint64 rowsDeleted = 1;</code>
-       */
-      public Builder setRowsDeleted(long value) {
-        bitField0_ |= 0x00000001;
-        rowsDeleted_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required uint64 rowsDeleted = 1;</code>
-       */
-      public Builder clearRowsDeleted() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        rowsDeleted_ = 0L;
-        onChanged();
-        return this;
-      }
-
-      // optional uint64 versionsDeleted = 2;
-      private long versionsDeleted_ ;
-      /**
-       * <code>optional uint64 versionsDeleted = 2;</code>
-       */
-      public boolean hasVersionsDeleted() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional uint64 versionsDeleted = 2;</code>
-       */
-      public long getVersionsDeleted() {
-        return versionsDeleted_;
-      }
-      /**
-       * <code>optional uint64 versionsDeleted = 2;</code>
-       */
-      public Builder setVersionsDeleted(long value) {
-        bitField0_ |= 0x00000002;
-        versionsDeleted_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional uint64 versionsDeleted = 2;</code>
-       */
-      public Builder clearVersionsDeleted() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        versionsDeleted_ = 0L;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.BulkDeleteResponse)
-    }
-
-    static {
-      defaultInstance = new BulkDeleteResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.BulkDeleteResponse)
-  }
-
-  /**
-   * Protobuf service {@code hbase.pb.BulkDeleteService}
-   */
-  public static abstract class BulkDeleteService
-      implements com.google.protobuf.Service {
-    protected BulkDeleteService() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc delete(.hbase.pb.BulkDeleteRequest) returns (.hbase.pb.BulkDeleteResponse);</code>
-       */
-      public abstract void delete(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse> done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new BulkDeleteService() {
-        @java.lang.Override
-        public  void delete(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse> done) {
-          impl.delete(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.delete(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc delete(.hbase.pb.BulkDeleteRequest) returns (.hbase.pb.BulkDeleteResponse);</code>
-     */
-    public abstract void delete(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse> done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.delete(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteService implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void delete(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.class,
-            org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse delete(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request)
-          throws com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse delete(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.example.generated.BulkDeleteProtos.BulkDeleteResponse.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.BulkDeleteService)
-  }
-
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_hbase_pb_BulkDeleteRequest_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_hbase_pb_BulkDeleteRequest_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_hbase_pb_BulkDeleteResponse_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_hbase_pb_BulkDeleteResponse_fieldAccessorTable;
-
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\020BulkDelete.proto\022\010hbase.pb\032\014Client.pro" +
-      "to\"\322\001\n\021BulkDeleteRequest\022\034\n\004scan\030\001 \002(\0132\016" +
-      ".hbase.pb.Scan\022:\n\ndeleteType\030\002 \002(\0162&.hba" +
-      "se.pb.BulkDeleteRequest.DeleteType\022\021\n\tti" +
-      "mestamp\030\003 \001(\004\022\024\n\014rowBatchSize\030\004 \002(\r\":\n\nD" +
-      "eleteType\022\007\n\003ROW\020\000\022\n\n\006FAMILY\020\001\022\n\n\006COLUMN" +
-      "\020\002\022\013\n\007VERSION\020\003\"B\n\022BulkDeleteResponse\022\023\n" +
-      "\013rowsDeleted\030\001 \002(\004\022\027\n\017versionsDeleted\030\002 " +
-      "\001(\0042X\n\021BulkDeleteService\022C\n\006delete\022\033.hba" +
-      "se.pb.BulkDeleteRequest\032\034.hbase.pb.BulkD",
-      "eleteResponseBQ\n5org.apache.hadoop.hbase" +
-      ".coprocessor.example.generatedB\020BulkDele" +
-      "teProtosH\001\210\001\001\240\001\001"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_hbase_pb_BulkDeleteRequest_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_hbase_pb_BulkDeleteRequest_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_pb_BulkDeleteRequest_descriptor,
-              new java.lang.String[] { "Scan", "DeleteType", "Timestamp", "RowBatchSize", });
-          internal_static_hbase_pb_BulkDeleteResponse_descriptor =
-            getDescriptor().getMessageTypes().get(1);
-          internal_static_hbase_pb_BulkDeleteResponse_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_pb_BulkDeleteResponse_descriptor,
-              new java.lang.String[] { "RowsDeleted", "VersionsDeleted", });
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(),
-        }, assigner);
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[27/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatValue.java
deleted file mode 100644
index f3315c8..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatValue.java
+++ /dev/null
@@ -1,454 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `float`.
- * The JSON representation for `FloatValue` is JSON number.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.FloatValue}
- */
-public  final class FloatValue extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.FloatValue)
-    FloatValueOrBuilder {
-  // Use FloatValue.newBuilder() to construct.
-  private FloatValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private FloatValue() {
-    value_ = 0F;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private FloatValue(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 13: {
-
-            value_ = input.readFloat();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_FloatValue_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_FloatValue_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private float value_;
-  /**
-   * <pre>
-   * The float value.
-   * </pre>
-   *
-   * <code>float value = 1;</code>
-   */
-  public float getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (value_ != 0F) {
-      output.writeFloat(1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (value_ != 0F) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeFloatSize(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue) obj;
-
-    boolean result = true;
-    result = result && (
-        java.lang.Float.floatToIntBits(getValue())
-        == java.lang.Float.floatToIntBits(
-            other.getValue()));
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + java.lang.Float.floatToIntBits(
-        getValue());
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `float`.
-   * The JSON representation for `FloatValue` is JSON number.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.FloatValue}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.FloatValue)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_FloatValue_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_FloatValue_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = 0F;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_FloatValue_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue.getDefaultInstance()) return this;
-      if (other.getValue() != 0F) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private float value_ ;
-    /**
-     * <pre>
-     * The float value.
-     * </pre>
-     *
-     * <code>float value = 1;</code>
-     */
-    public float getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * The float value.
-     * </pre>
-     *
-     * <code>float value = 1;</code>
-     */
-    public Builder setValue(float value) {
-      
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The float value.
-     * </pre>
-     *
-     * <code>float value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = 0F;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.FloatValue)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.FloatValue)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FloatValue>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<FloatValue>() {
-    public FloatValue parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new FloatValue(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FloatValue> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FloatValue> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.FloatValue getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatValueOrBuilder.java
deleted file mode 100644
index a5302d7..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface FloatValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.FloatValue)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The float value.
-   * </pre>
-   *
-   * <code>float value = 1;</code>
-   */
-  float getValue();
-}


[04/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java
deleted file mode 100644
index 200a7f4..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestRpcServiceProtos.java
+++ /dev/null
@@ -1,575 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: test_rpc_service.proto
-
-package org.apache.hadoop.hbase.shaded.ipc.protobuf.generated;
-
-public final class TestRpcServiceProtos {
-  private TestRpcServiceProtos() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  /**
-   * <pre>
-   **
-   * A protobuf service for use in tests
-   * </pre>
-   *
-   * Protobuf service {@code TestProtobufRpcProto}
-   */
-  public static abstract class TestProtobufRpcProto
-      implements org.apache.hadoop.hbase.shaded.com.google.protobuf.Service {
-    protected TestProtobufRpcProto() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
-       */
-      public abstract void ping(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
-      /**
-       * <code>rpc echo(.EchoRequestProto) returns (.EchoResponseProto);</code>
-       */
-      public abstract void echo(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done);
-
-      /**
-       * <code>rpc error(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
-       */
-      public abstract void error(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
-      /**
-       * <code>rpc pause(.PauseRequestProto) returns (.EmptyResponseProto);</code>
-       */
-      public abstract void pause(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
-      /**
-       * <code>rpc addr(.EmptyRequestProto) returns (.AddrResponseProto);</code>
-       */
-      public abstract void addr(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done);
-
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new TestProtobufRpcProto() {
-        @java.lang.Override
-        public  void ping(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
-          impl.ping(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void echo(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done) {
-          impl.echo(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void error(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
-          impl.error(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void pause(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
-          impl.pause(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void addr(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done) {
-          impl.addr(controller, request, done);
-        }
-
-      };
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingService() {
-        public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message callBlockingMethod(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Message request)
-            throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.ping(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request);
-            case 1:
-              return impl.echo(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto)request);
-            case 2:
-              return impl.error(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request);
-            case 3:
-              return impl.pause(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto)request);
-            case 4:
-              return impl.addr(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
-            getRequestPrototype(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.getDefaultInstance();
-            case 2:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-            case 3:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto.getDefaultInstance();
-            case 4:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
-            getResponsePrototype(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance();
-            case 2:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-            case 3:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-            case 4:
-              return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
-     */
-    public abstract void ping(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
-    /**
-     * <code>rpc echo(.EchoRequestProto) returns (.EchoResponseProto);</code>
-     */
-    public abstract void echo(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done);
-
-    /**
-     * <code>rpc error(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
-     */
-    public abstract void error(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
-    /**
-     * <code>rpc pause(.PauseRequestProto) returns (.EmptyResponseProto);</code>
-     */
-    public abstract void pause(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done);
-
-    /**
-     * <code>rpc addr(.EmptyRequestProto) returns (.AddrResponseProto);</code>
-     */
-    public abstract void addr(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done);
-
-    public static final
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestRpcServiceProtos.getDescriptor().getServices().get(0);
-    }
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Message request,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.ping(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
-              done));
-          return;
-        case 1:
-          this.echo(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto)request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto>specializeCallback(
-              done));
-          return;
-        case 2:
-          this.error(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
-              done));
-          return;
-        case 3:
-          this.pause(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto)request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
-              done));
-          return;
-        case 4:
-          this.addr(controller, (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
-        getRequestPrototype(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto.getDefaultInstance();
-        case 2:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-        case 3:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto.getDefaultInstance();
-        case 4:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
-        getResponsePrototype(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance();
-        case 2:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-        case 3:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-        case 4:
-          return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestRpcServiceProtos.TestProtobufRpcProto implements Interface {
-      private Stub(org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcChannel channel;
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void ping(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
-      }
-
-      public  void echo(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance(),
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.class,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance()));
-      }
-
-      public  void error(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(2),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
-      }
-
-      public  void pause(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(3),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
-      }
-
-      public  void addr(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(4),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.getDefaultInstance(),
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.class,
-            org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto ping(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto echo(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto error(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto pause(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto addr(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final org.apache.hadoop.hbase.shaded.com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto ping(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto echo(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto error(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(2),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto pause(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.PauseRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(3),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto addr(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(4),
-          controller,
-          request,
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.AddrResponseProto.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:TestProtobufRpcProto)
-  }
-
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\026test_rpc_service.proto\032\ntest.proto2\212\002\n" +
-      "\024TestProtobufRpcProto\022/\n\004ping\022\022.EmptyReq" +
-      "uestProto\032\023.EmptyResponseProto\022-\n\004echo\022\021" +
-      ".EchoRequestProto\032\022.EchoResponseProto\0220\n" +
-      "\005error\022\022.EmptyRequestProto\032\023.EmptyRespon" +
-      "seProto\0220\n\005pause\022\022.PauseRequestProto\032\023.E" +
-      "mptyResponseProto\022.\n\004addr\022\022.EmptyRequest" +
-      "Proto\032\022.AddrResponseProtoBS\n5org.apache." +
-      "hadoop.hbase.shaded.ipc.protobuf.generat" +
-      "edB\024TestRpcServiceProtos\210\001\001\240\001\001"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-          org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.getDescriptor(),
-        }, assigner);
-    org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProtos.getDescriptor();
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[10/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnknownFieldSetLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnknownFieldSetLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnknownFieldSetLite.java
deleted file mode 100644
index faa4c30..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnknownFieldSetLite.java
+++ /dev/null
@@ -1,432 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.util.Arrays;
-
-/**
- * {@code UnknownFieldSetLite} is used to keep track of fields which were seen
- * when parsing a protocol message but whose field numbers or types are
- * unrecognized. This most frequently occurs when new fields are added to a
- * message type and then messages containing those fields are read by old
- * software that was compiled before the new types were added.
- *
- * <p>For use by generated code only.
- *
- * @author dweis@google.com (Daniel Weis)
- */
-public final class UnknownFieldSetLite {
-  
-  // Arbitrarily chosen.
-  // TODO(dweis): Tune this number?
-  private static final int MIN_CAPACITY = 8;
-
-  private static final UnknownFieldSetLite DEFAULT_INSTANCE =
-      new UnknownFieldSetLite(0, new int[0], new Object[0], false /* isMutable */);
-
-  /**
-   * Get an empty {@code UnknownFieldSetLite}.
-   *
-   * <p>For use by generated code only.
-   */
-  public static UnknownFieldSetLite getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-  
-  /**
-   * Returns a new mutable instance.
-   */
-  static UnknownFieldSetLite newInstance() {
-    return new UnknownFieldSetLite();
-  }
-
-  /**
-   * Returns a mutable {@code UnknownFieldSetLite} that is the composite of {@code first} and
-   * {@code second}.
-   */
-  static UnknownFieldSetLite mutableCopyOf(UnknownFieldSetLite first, UnknownFieldSetLite second) {
-    int count = first.count + second.count;
-    int[] tags = Arrays.copyOf(first.tags, count);
-    System.arraycopy(second.tags, 0, tags, first.count, second.count);
-    Object[] objects = Arrays.copyOf(first.objects, count);
-    System.arraycopy(second.objects, 0, objects, first.count, second.count);
-    return new UnknownFieldSetLite(count, tags, objects, true /* isMutable */);
-  }
-  
-  /**
-   * The number of elements in the set.
-   */
-  private int count;
-  
-  /**
-   * The tag numbers for the elements in the set.
-   */
-  private int[] tags;
-  
-  /**
-   * The boxed values of the elements in the set.
-   */
-  private Object[] objects;
-  
-  /**
-   * The lazily computed serialized size of the set.
-   */
-  private int memoizedSerializedSize = -1;
-  
-  /**
-   * Indicates that this object is mutable. 
-   */
-  private boolean isMutable;
-
-  /**
-   * Constructs a mutable {@code UnknownFieldSetLite}.
-   */
-  private UnknownFieldSetLite() {
-    this(0, new int[MIN_CAPACITY], new Object[MIN_CAPACITY], true /* isMutable */);
-  }
-  
-  /**
-   * Constructs the {@code UnknownFieldSetLite}.
-   */
-  private UnknownFieldSetLite(int count, int[] tags, Object[] objects, boolean isMutable) {
-    this.count = count;
-    this.tags = tags;
-    this.objects = objects;
-    this.isMutable = isMutable;
-  }
-  
-  /**
-   * Marks this object as immutable.
-   * 
-   * <p>Future calls to methods that attempt to modify this object will throw.
-   */
-  public void makeImmutable() {
-    this.isMutable = false;
-  }
-  
-  /**
-   * Throws an {@link UnsupportedOperationException} if immutable.
-   */
-  void checkMutable() {
-    if (!isMutable) {
-      throw new UnsupportedOperationException(); 
-    }
-  }
-
-  /**
-   * Serializes the set and writes it to {@code output}.
-   *
-   * <p>For use by generated code only.
-   */
-  public void writeTo(CodedOutputStream output) throws IOException {
-    for (int i = 0; i < count; i++) {
-      int tag = tags[i];
-      int fieldNumber = WireFormat.getTagFieldNumber(tag);
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          output.writeUInt64(fieldNumber, (Long) objects[i]);
-          break;
-        case WireFormat.WIRETYPE_FIXED32:
-          output.writeFixed32(fieldNumber, (Integer) objects[i]);
-          break;
-        case WireFormat.WIRETYPE_FIXED64:
-          output.writeFixed64(fieldNumber, (Long) objects[i]);
-          break;
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          output.writeBytes(fieldNumber, (ByteString) objects[i]);
-          break;
-        case WireFormat.WIRETYPE_START_GROUP:
-          output.writeTag(fieldNumber, WireFormat.WIRETYPE_START_GROUP);
-          ((UnknownFieldSetLite) objects[i]).writeTo(output);
-          output.writeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP);
-          break;
-        default:
-          throw InvalidProtocolBufferException.invalidWireType();
-      }
-    }
-  }
-
-  /**
-   * Serializes the set and writes it to {@code output} using {@code MessageSet} wire format.
-   *
-   * <p>For use by generated code only.
-   */
-  public void writeAsMessageSetTo(CodedOutputStream output) throws IOException {
-    for (int i = 0; i < count; i++) {
-      int fieldNumber = WireFormat.getTagFieldNumber(tags[i]);
-      output.writeRawMessageSetExtension(fieldNumber, (ByteString) objects[i]);
-    }
-  }
-
-
-  /**
-   * Get the number of bytes required to encode this field, including field number, using {@code
-   * MessageSet} wire format.
-   */
-  public int getSerializedSizeAsMessageSet() {
-    int size = memoizedSerializedSize;
-    if (size != -1) {
-      return size;
-    }
-    
-    size = 0;
-    for (int i = 0; i < count; i++) {
-      int tag = tags[i];
-      int fieldNumber = WireFormat.getTagFieldNumber(tag);
-      size += CodedOutputStream.computeRawMessageSetExtensionSize(
-          fieldNumber, (ByteString) objects[i]);
-    }
-    
-    memoizedSerializedSize = size;
-    
-    return size;
-  }
-
-  /**
-   * Get the number of bytes required to encode this set.
-   *
-   * <p>For use by generated code only.
-   */
-  public int getSerializedSize() {
-    int size = memoizedSerializedSize;
-    if (size != -1) {
-      return size;
-    }
-    
-    size = 0;
-    for (int i = 0; i < count; i++) {
-      int tag = tags[i];
-      int fieldNumber = WireFormat.getTagFieldNumber(tag);
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          size += CodedOutputStream.computeUInt64Size(fieldNumber, (Long) objects[i]);
-          break;
-        case WireFormat.WIRETYPE_FIXED32:
-          size += CodedOutputStream.computeFixed32Size(fieldNumber, (Integer) objects[i]);
-          break;
-        case WireFormat.WIRETYPE_FIXED64:
-          size += CodedOutputStream.computeFixed64Size(fieldNumber, (Long) objects[i]);
-          break;
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          size += CodedOutputStream.computeBytesSize(fieldNumber, (ByteString) objects[i]);
-          break;
-        case WireFormat.WIRETYPE_START_GROUP:
-          size +=  CodedOutputStream.computeTagSize(fieldNumber) * 2
-              + ((UnknownFieldSetLite) objects[i]).getSerializedSize();
-          break;
-        default:
-          throw new IllegalStateException(InvalidProtocolBufferException.invalidWireType());
-      }
-    }
-    
-    memoizedSerializedSize = size;
-    
-    return size;
-  }
-  
-  private static boolean equals(int[] tags1, int[] tags2, int count) {
-    for (int i = 0; i < count; ++i) {
-      if (tags1[i] != tags2[i]) {
-        return false;
-      }
-    }
-    return true;
-  }
-
-  private static boolean equals(Object[] objects1, Object[] objects2, int count) {
-    for (int i = 0; i < count; ++i) {
-      if (!objects1[i].equals(objects2[i])) {
-        return false;
-      }
-    }
-    return true;
-  }
-
-  @Override
-  public boolean equals(Object obj) {
-    if (this == obj) {
-      return true;
-    }
-
-    if (obj == null) {
-      return false;
-    }
-
-    if (!(obj instanceof UnknownFieldSetLite)) {
-      return false;
-    }
-    
-    UnknownFieldSetLite other = (UnknownFieldSetLite) obj;    
-    if (count != other.count
-        || !equals(tags, other.tags, count)
-        || !equals(objects, other.objects, count)) {
-      return false;
-    }
-
-    return true;
-  }
-
-  @Override
-  public int hashCode() {
-    int hashCode = 17;
-    
-    hashCode = 31 * hashCode + count;
-    hashCode = 31 * hashCode + Arrays.hashCode(tags);
-    hashCode = 31 * hashCode + Arrays.deepHashCode(objects);
-    
-    return hashCode;
-  }
-
-  /**
-   * Prints a String representation of the unknown field set.
-   *
-   * <p>For use by generated code only.
-   *
-   * @param buffer the buffer to write to
-   * @param indent the number of spaces the fields should be indented by
-   */
-  final void printWithIndent(StringBuilder buffer, int indent) {
-    for (int i = 0; i < count; i++) {
-      int fieldNumber = WireFormat.getTagFieldNumber(tags[i]);
-      MessageLiteToString.printField(buffer, indent, String.valueOf(fieldNumber), objects[i]);
-    }
-  }
-
-  // Package private for unsafe experimental runtime.
-  void storeField(int tag, Object value) {
-    ensureCapacity();
-    
-    tags[count] = tag;
-    objects[count] = value;
-    count++;
-  }
-  
-  /**
-   * Ensures that our arrays are long enough to store more metadata.
-   */
-  private void ensureCapacity() {
-    if (count == tags.length) {        
-      int increment = count < (MIN_CAPACITY / 2) ? MIN_CAPACITY : count >> 1;
-      int newLength = count + increment;
-        
-      tags = Arrays.copyOf(tags, newLength);
-      objects = Arrays.copyOf(objects, newLength);
-    }
-  }
-  
-  /**
-   * Parse a single field from {@code input} and merge it into this set.
-   *
-   * <p>For use by generated code only.
-   *
-   * @param tag The field's tag number, which was already parsed.
-   * @return {@code false} if the tag is an end group tag.
-   */
-  boolean mergeFieldFrom(final int tag, final CodedInputStream input) throws IOException {
-    checkMutable();
-    final int fieldNumber = WireFormat.getTagFieldNumber(tag);
-    switch (WireFormat.getTagWireType(tag)) {
-      case WireFormat.WIRETYPE_VARINT:
-        storeField(tag, input.readInt64());
-        return true;
-      case WireFormat.WIRETYPE_FIXED32:
-        storeField(tag, input.readFixed32());
-        return true;
-      case WireFormat.WIRETYPE_FIXED64:
-        storeField(tag, input.readFixed64());
-        return true;
-      case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-        storeField(tag, input.readBytes());
-        return true;
-      case WireFormat.WIRETYPE_START_GROUP:
-        final UnknownFieldSetLite subFieldSet = new UnknownFieldSetLite();
-        subFieldSet.mergeFrom(input);
-        input.checkLastTagWas(
-            WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP));
-        storeField(tag, subFieldSet);
-        return true;
-      case WireFormat.WIRETYPE_END_GROUP:
-        return false;
-      default:
-        throw InvalidProtocolBufferException.invalidWireType();
-    }
-  }
-
-  /**
-   * Convenience method for merging a new field containing a single varint
-   * value. This is used in particular when an unknown enum value is
-   * encountered.
-   *
-   * <p>For use by generated code only.
-   */
-  UnknownFieldSetLite mergeVarintField(int fieldNumber, int value) {
-    checkMutable();
-    if (fieldNumber == 0) {
-      throw new IllegalArgumentException("Zero is not a valid field number.");
-    }
-
-    storeField(WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_VARINT), (long) value);
-    
-    return this;
-  }
-
-  /**
-   * Convenience method for merging a length-delimited field.
-   *
-   * <p>For use by generated code only.
-   */
-  UnknownFieldSetLite mergeLengthDelimitedField(final int fieldNumber, final ByteString value) {  
-    checkMutable();
-    if (fieldNumber == 0) {
-      throw new IllegalArgumentException("Zero is not a valid field number.");
-    }
-
-    storeField(WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_LENGTH_DELIMITED), value);
-    
-    return this;
-  }
-  
-  /**
-   * Parse an entire message from {@code input} and merge its fields into
-   * this set.
-   */
-  private UnknownFieldSetLite mergeFrom(final CodedInputStream input) throws IOException {
-    // Ensures initialization in mergeFieldFrom.
-    while (true) {
-      final int tag = input.readTag();
-      if (tag == 0 || !mergeFieldFrom(tag, input)) {
-        break;
-      }
-    }
-    return this;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnmodifiableLazyStringList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnmodifiableLazyStringList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnmodifiableLazyStringList.java
deleted file mode 100644
index 54786dd..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnmodifiableLazyStringList.java
+++ /dev/null
@@ -1,210 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.AbstractList;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.List;
-import java.util.ListIterator;
-import java.util.RandomAccess;
-
-/**
- * An implementation of {@link LazyStringList} that wraps another
- * {@link LazyStringList} such that it cannot be modified via the wrapper.
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public class UnmodifiableLazyStringList extends AbstractList<String>
-    implements LazyStringList, RandomAccess {
-
-  private final LazyStringList list;
-
-  public UnmodifiableLazyStringList(LazyStringList list) {
-    this.list = list;
-  }
-
-  @Override
-  public String get(int index) {
-    return list.get(index);
-  }
-  
-  @Override
-  public Object getRaw(int index) {
-    return list.getRaw(index);
-  }
-
-  @Override
-  public int size() {
-    return list.size();
-  }
-
-  @Override
-  public ByteString getByteString(int index) {
-    return list.getByteString(index);
-  }
-
-  @Override
-  public void add(ByteString element) {
-    throw new UnsupportedOperationException();
-  }
-
-  @Override
-  public void set(int index, ByteString element) {
-    throw new UnsupportedOperationException();
-  }
-
-  @Override
-  public boolean addAllByteString(Collection<? extends ByteString> element) {
-    throw new UnsupportedOperationException();
-  }
-
-  @Override
-  public byte[] getByteArray(int index) {
-    return list.getByteArray(index);
-  }
-
-  @Override
-  public void add(byte[] element) {
-    throw new UnsupportedOperationException();
-  }
-  
-  @Override
-  public void set(int index, byte[] element) {
-    throw new UnsupportedOperationException();
-  }
-
-  @Override
-  public boolean addAllByteArray(Collection<byte[]> element) {
-    throw new UnsupportedOperationException();
-  }
-
-  @Override
-  public ListIterator<String> listIterator(final int index) {
-    return new ListIterator<String>() {
-      ListIterator<String> iter = list.listIterator(index);
-
-      @Override
-      public boolean hasNext() {
-        return iter.hasNext();
-      }
-
-      @Override
-      public String next() {
-        return iter.next();
-      }
-
-      @Override
-      public boolean hasPrevious() {
-        return iter.hasPrevious();
-      }
-
-      @Override
-      public String previous() {
-        return iter.previous();
-      }
-
-      @Override
-      public int nextIndex() {
-        return iter.nextIndex();
-      }
-
-      @Override
-      public int previousIndex() {
-        return iter.previousIndex();
-      }
-
-      @Override
-      public void remove() {
-        throw new UnsupportedOperationException();
-      }
-
-      @Override
-      public void set(String o) {
-        throw new UnsupportedOperationException();
-      }
-
-      @Override
-      public void add(String o) {
-        throw new UnsupportedOperationException();
-      }
-    };
-  }
-
-  @Override
-  public Iterator<String> iterator() {
-    return new Iterator<String>() {
-      Iterator<String> iter = list.iterator();
-
-      @Override
-      public boolean hasNext() {
-        return iter.hasNext();
-      }
-
-      @Override
-      public String next() {
-        return iter.next();
-      }
-
-      @Override
-      public void remove() {
-        throw new UnsupportedOperationException();
-      }
-    };
-  }
-
-  @Override
-  public List<?> getUnderlyingElements() {
-    // The returned value is already unmodifiable.
-    return list.getUnderlyingElements();
-  }
-
-  @Override
-  public void mergeFrom(LazyStringList other) {
-    throw new UnsupportedOperationException();
-  }
-
-  @Override
-  public List<byte[]> asByteArrayList() {
-    return Collections.unmodifiableList(list.asByteArrayList());
-  }
-
-  @Override
-  public List<ByteString> asByteStringList() {
-    return Collections.unmodifiableList(list.asByteStringList());
-  }
-
-  @Override
-  public LazyStringList getUnmodifiableView() {
-    return this;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnsafeByteOperations.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnsafeByteOperations.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnsafeByteOperations.java
deleted file mode 100644
index ad99372..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnsafeByteOperations.java
+++ /dev/null
@@ -1,131 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.nio.ByteBuffer;
-
-/**
- * Provides a number of unsafe byte operations to be used by advanced applications with high
- * performance requirements. These methods are referred to as "unsafe" due to the fact that they
- * potentially expose the backing buffer of a {@link ByteString} to the application.
- *
- * <p><strong>DISCLAIMER:</strong> The methods in this class should only be called if it is
- * guaranteed that the buffer backing the {@link ByteString} will never change! Mutation of a
- * {@link ByteString} can lead to unexpected and undesirable consequences in your application,
- * and will likely be difficult to debug. Proceed with caution!
- *
- * <p>This can have a number of significant side affects that have
- * spooky-action-at-a-distance-like behavior. In particular, if the bytes value changes out from
- * under a Protocol Buffer:
- * <ul>
- * <li>serialization may throw
- * <li>serialization may succeed but the wrong bytes may be written out
- * <li>messages are no longer threadsafe
- * <li>hashCode may be incorrect
- *   <ul>
- *   <li>can result in a permanent memory leak when used as a key in a long-lived HashMap
- *   <li> the semantics of many programs may be violated if this is the case
- *   </ul>
- * </ul>
- * Each of these issues will occur in parts of the code base that are entirely distinct from the
- * parts of the code base modifying the buffer. In fact, both parts of the code base may be correct
- * - it is the bridging with the unsafe operations that was in error!
- */
-@ExperimentalApi
-public final class UnsafeByteOperations {
-  private UnsafeByteOperations() {}
-
-  /**
-   * An unsafe operation that returns a {@link ByteString} that is backed by the provided buffer.
-   *
-   * @param buffer the buffer to be wrapped
-   * @return a {@link ByteString} backed by the provided buffer
-   */
-  public static ByteString unsafeWrap(byte[] buffer) {
-    return ByteString.wrap(buffer);
-  }
-
-  /**
-   * An unsafe operation that returns a {@link ByteString} that is backed by a subregion of the
-   * provided buffer.
-   *
-   * @param buffer the buffer to be wrapped
-   * @param offset the offset of the wrapped region
-   * @param length the number of bytes of the wrapped region
-   * @return a {@link ByteString} backed by the provided buffer
-   */
-  public static ByteString unsafeWrap(byte[] buffer, int offset, int length) {
-    return ByteString.wrap(buffer, offset, length);
-  }
-
-  /**
-   * An unsafe operation that returns a {@link ByteString} that is backed by the provided buffer.
-   *
-   * @param buffer the Java NIO buffer to be wrapped
-   * @return a {@link ByteString} backed by the provided buffer
-   */
-  public static ByteString unsafeWrap(ByteBuffer buffer) {
-    return ByteString.wrap(buffer);
-  }
-
-  /**
-   * An unsafe operation that returns a {@link ByteString} that is backed by the provided buffer.
-   * @param buffer the ByteInput buffer to be wrapped
-   * @param offset the offset of the wrapped byteinput
-   * @param length the number of bytes of the byteinput
-   * @return a {@link ByteString} backed by the provided buffer
-   */
-  public static ByteString unsafeWrap(ByteInput buffer, int offset, int len) {
-    return ByteString.wrap(buffer, offset, len);
-  }
-
-  /**
-   * Writes the given {@link ByteString} to the provided {@link ByteOutput}. Calling this method may
-   * result in multiple operations on the target {@link ByteOutput}
-   * (i.e. for roped {@link ByteString}s).
-   *
-   * <p>This method exposes the internal backing buffer(s) of the {@link ByteString} to the {@link
-   * ByteOutput} in order to avoid additional copying overhead. It would be possible for a malicious
-   * {@link ByteOutput} to corrupt the {@link ByteString}. Use with caution!
-   *
-   * <p> NOTE: The {@link ByteOutput} <strong>MUST NOT</strong> modify the provided buffers. Doing
-   * so may result in corrupted data, which would be difficult to debug.
-   *
-   * @param bytes the {@link ByteString} to be written
-   * @param  output  the output to receive the bytes
-   * @throws IOException  if an I/O error occurs
-   */
-  public static void unsafeWriteTo(ByteString bytes, ByteOutput output) throws IOException {
-    bytes.writeTo(output);
-  }
-
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnsafeUtil.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnsafeUtil.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnsafeUtil.java
deleted file mode 100644
index bb4d8d6..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnsafeUtil.java
+++ /dev/null
@@ -1,295 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.lang.reflect.Field;
-import java.nio.Buffer;
-import java.nio.ByteBuffer;
-import java.security.AccessController;
-import java.security.PrivilegedExceptionAction;
-import sun.misc.Unsafe;
-
-/** Utility class for working with unsafe operations. */
-// TODO(nathanmittler): Add support for Android Memory/MemoryBlock
-final class UnsafeUtil {
-  private static final sun.misc.Unsafe UNSAFE = getUnsafe();
-  private static final boolean HAS_UNSAFE_BYTEBUFFER_OPERATIONS =
-      supportsUnsafeByteBufferOperations();
-  private static final boolean HAS_UNSAFE_ARRAY_OPERATIONS = supportsUnsafeArrayOperations();
-  private static final long ARRAY_BASE_OFFSET = byteArrayBaseOffset();
-  private static final long BUFFER_ADDRESS_OFFSET = fieldOffset(field(Buffer.class, "address"));
-
-  private UnsafeUtil() {}
-
-  static boolean hasUnsafeArrayOperations() {
-    return HAS_UNSAFE_ARRAY_OPERATIONS;
-  }
-
-  static boolean hasUnsafeByteBufferOperations() {
-    return HAS_UNSAFE_BYTEBUFFER_OPERATIONS;
-  }
-
-  static Object allocateInstance(Class<?> clazz) {
-    try {
-      return UNSAFE.allocateInstance(clazz);
-    } catch (InstantiationException e) {
-      throw new RuntimeException(e);
-    }
-  }
-
-  static long objectFieldOffset(Field field) {
-    return UNSAFE.objectFieldOffset(field);
-  }
-
-  static long getArrayBaseOffset() {
-    return ARRAY_BASE_OFFSET;
-  }
-
-  static byte getByte(Object target, long offset) {
-    return UNSAFE.getByte(target, offset);
-  }
-
-  static void putByte(Object target, long offset, byte value) {
-    UNSAFE.putByte(target, offset, value);
-  }
-
-  static int getInt(Object target, long offset) {
-    return UNSAFE.getInt(target, offset);
-  }
-
-  static void putInt(Object target, long offset, int value) {
-    UNSAFE.putInt(target, offset, value);
-  }
-
-  static long getLong(Object target, long offset) {
-    return UNSAFE.getLong(target, offset);
-  }
-
-  static void putLong(Object target, long offset, long value) {
-    UNSAFE.putLong(target, offset, value);
-  }
-
-  static boolean getBoolean(Object target, long offset) {
-    return UNSAFE.getBoolean(target, offset);
-  }
-
-  static void putBoolean(Object target, long offset, boolean value) {
-    UNSAFE.putBoolean(target, offset, value);
-  }
-
-  static float getFloat(Object target, long offset) {
-    return UNSAFE.getFloat(target, offset);
-  }
-
-  static void putFloat(Object target, long offset, float value) {
-    UNSAFE.putFloat(target, offset, value);
-  }
-
-  static double getDouble(Object target, long offset) {
-    return UNSAFE.getDouble(target, offset);
-  }
-
-  static void putDouble(Object target, long offset, double value) {
-    UNSAFE.putDouble(target, offset, value);
-  }
-
-  static Object getObject(Object target, long offset) {
-    return UNSAFE.getObject(target, offset);
-  }
-
-  static void putObject(Object target, long offset, Object value) {
-    UNSAFE.putObject(target, offset, value);
-  }
-
-  static void copyMemory(
-      Object src, long srcOffset, Object target, long targetOffset, long length) {
-    UNSAFE.copyMemory(src, srcOffset, target, targetOffset, length);
-  }
-
-  static byte getByte(long address) {
-    return UNSAFE.getByte(address);
-  }
-
-  static void putByte(long address, byte value) {
-    UNSAFE.putByte(address, value);
-  }
-
-  static int getInt(long address) {
-    return UNSAFE.getInt(address);
-  }
-
-  static void putInt(long address, int value) {
-    UNSAFE.putInt(address, value);
-  }
-
-  static long getLong(long address) {
-    return UNSAFE.getLong(address);
-  }
-
-  static void putLong(long address, long value) {
-    UNSAFE.putLong(address, value);
-  }
-
-  static void copyMemory(long srcAddress, long targetAddress, long length) {
-    UNSAFE.copyMemory(srcAddress, targetAddress, length);
-  }
-
-  static void setMemory(long address, long numBytes, byte value) {
-    UNSAFE.setMemory(address, numBytes, value);
-  }
-
-  /**
-   * Gets the offset of the {@code address} field of the given direct {@link ByteBuffer}.
-   */
-  static long addressOffset(ByteBuffer buffer) {
-    return UNSAFE.getLong(buffer, BUFFER_ADDRESS_OFFSET);
-  }
-
-  /**
-   * Gets the {@code sun.misc.Unsafe} instance, or {@code null} if not available on this platform.
-   */
-  private static sun.misc.Unsafe getUnsafe() {
-    sun.misc.Unsafe unsafe = null;
-    try {
-      unsafe =
-          AccessController.doPrivileged(
-              new PrivilegedExceptionAction<Unsafe>() {
-                @Override
-                public sun.misc.Unsafe run() throws Exception {
-                  Class<sun.misc.Unsafe> k = sun.misc.Unsafe.class;
-
-                  for (Field f : k.getDeclaredFields()) {
-                    f.setAccessible(true);
-                    Object x = f.get(null);
-                    if (k.isInstance(x)) {
-                      return k.cast(x);
-                    }
-                  }
-                  // The sun.misc.Unsafe field does not exist.
-                  return null;
-                }
-              });
-    } catch (Throwable e) {
-      // Catching Throwable here due to the fact that Google AppEngine raises NoClassDefFoundError
-      // for Unsafe.
-    }
-    return unsafe;
-  }
-
-  /** Indicates whether or not unsafe array operations are supported on this platform. */
-  private static boolean supportsUnsafeArrayOperations() {
-    boolean supported = false;
-    if (UNSAFE != null) {
-      try {
-        Class<?> clazz = UNSAFE.getClass();
-        clazz.getMethod("objectFieldOffset", Field.class);
-        clazz.getMethod("allocateInstance", Class.class);
-        clazz.getMethod("arrayBaseOffset", Class.class);
-        clazz.getMethod("getByte", Object.class, long.class);
-        clazz.getMethod("putByte", Object.class, long.class, byte.class);
-        clazz.getMethod("getBoolean", Object.class, long.class);
-        clazz.getMethod("putBoolean", Object.class, long.class, boolean.class);
-        clazz.getMethod("getInt", Object.class, long.class);
-        clazz.getMethod("putInt", Object.class, long.class, int.class);
-        clazz.getMethod("getLong", Object.class, long.class);
-        clazz.getMethod("putLong", Object.class, long.class, long.class);
-        clazz.getMethod("getFloat", Object.class, long.class);
-        clazz.getMethod("putFloat", Object.class, long.class, float.class);
-        clazz.getMethod("getDouble", Object.class, long.class);
-        clazz.getMethod("putDouble", Object.class, long.class, double.class);
-        clazz.getMethod("getObject", Object.class, long.class);
-        clazz.getMethod("putObject", Object.class, long.class, Object.class);
-        clazz.getMethod(
-            "copyMemory", Object.class, long.class, Object.class, long.class, long.class);
-        supported = true;
-      } catch (Throwable e) {
-        // Do nothing.
-      }
-    }
-    return supported;
-  }
-
-  private static boolean supportsUnsafeByteBufferOperations() {
-    boolean supported = false;
-    if (UNSAFE != null) {
-      try {
-        Class<?> clazz = UNSAFE.getClass();
-        // Methods for getting direct buffer address.
-        clazz.getMethod("objectFieldOffset", Field.class);
-        clazz.getMethod("getLong", Object.class, long.class);
-
-        clazz.getMethod("getByte", long.class);
-        clazz.getMethod("putByte", long.class, byte.class);
-        clazz.getMethod("getInt", long.class);
-        clazz.getMethod("putInt", long.class, int.class);
-        clazz.getMethod("getLong", long.class);
-        clazz.getMethod("putLong", long.class, long.class);
-        clazz.getMethod("setMemory", long.class, long.class, byte.class);
-        clazz.getMethod("copyMemory", long.class, long.class, long.class);
-        supported = true;
-      } catch (Throwable e) {
-        // Do nothing.
-      }
-    }
-    return supported;
-  }
-
-  /**
-   * Get the base offset for byte arrays, or {@code -1} if {@code sun.misc.Unsafe} is not available.
-   */
-  private static int byteArrayBaseOffset() {
-    return HAS_UNSAFE_ARRAY_OPERATIONS ? UNSAFE.arrayBaseOffset(byte[].class) : -1;
-  }
-
-  /**
-   * Returns the offset of the provided field, or {@code -1} if {@code sun.misc.Unsafe} is not
-   * available.
-   */
-  private static long fieldOffset(Field field) {
-    return field == null || UNSAFE == null ? -1 : UNSAFE.objectFieldOffset(field);
-  }
-
-  /**
-   * Gets the field with the given name within the class, or {@code null} if not found. If found,
-   * the field is made accessible.
-   */
-  private static Field field(Class<?> clazz, String fieldName) {
-    Field field;
-    try {
-      field = clazz.getDeclaredField(fieldName);
-      field.setAccessible(true);
-    } catch (Throwable t) {
-      // Failed to access the fields.
-      field = null;
-    }
-    return field;
-  }
-}


[37/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteString.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteString.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteString.java
deleted file mode 100644
index f673ee3..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteString.java
+++ /dev/null
@@ -1,1565 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.ByteArrayInputStream;
-import java.io.ByteArrayOutputStream;
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.InvalidObjectException;
-import java.io.ObjectInputStream;
-import java.io.OutputStream;
-import java.io.Serializable;
-import java.io.UnsupportedEncodingException;
-import java.nio.ByteBuffer;
-import java.nio.charset.Charset;
-import java.nio.charset.UnsupportedCharsetException;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.List;
-import java.util.NoSuchElementException;
-
-/**
- * Immutable sequence of bytes. Substring is supported by sharing the reference to the immutable
- * underlying bytes. Concatenation is likewise supported without copying (long strings) by building
- * a tree of pieces in {@link RopeByteString}.
- *
- * <p>Like {@link String}, the contents of a {@link ByteString} can never be observed to change, not
- * even in the presence of a data race or incorrect API usage in the client code.
- *
- * @author crazybob@google.com Bob Lee
- * @author kenton@google.com Kenton Varda
- * @author carlanton@google.com Carl Haverl
- * @author martinrb@google.com Martin Buchholz
- */
-public abstract class ByteString implements Iterable<Byte>, Serializable {
-
-  /**
-   * When two strings to be concatenated have a combined length shorter than
-   * this, we just copy their bytes on {@link #concat(ByteString)}.
-   * The trade-off is copy size versus the overhead of creating tree nodes
-   * in {@link RopeByteString}.
-   */
-  static final int CONCATENATE_BY_COPY_SIZE = 128;
-
-  /**
-   * When copying an InputStream into a ByteString with .readFrom(),
-   * the chunks in the underlying rope start at 256 bytes, but double
-   * each iteration up to 8192 bytes.
-   */
-  static final int MIN_READ_FROM_CHUNK_SIZE = 0x100;  // 256b
-  static final int MAX_READ_FROM_CHUNK_SIZE = 0x2000;  // 8k
-
-  /**
-   * Empty {@code ByteString}.
-   */
-  public static final ByteString EMPTY = new LiteralByteString(Internal.EMPTY_BYTE_ARRAY);
-
-  /**
-   * An interface to efficiently copy {@code byte[]}.
-   *
-   * <p>One of the noticeable costs of copying a byte[] into a new array using
-   * {@code System.arraycopy} is nullification of a new buffer before the copy. It has been shown
-   * the Hotspot VM is capable to intrisicfy {@code Arrays.copyOfRange} operation to avoid this
-   * expensive nullification and provide substantial performance gain. Unfortunately this does not
-   * hold on Android runtimes and could make the copy slightly slower due to additional code in
-   * the {@code Arrays.copyOfRange}. Thus we provide two different implementation for array copier
-   * for Hotspot and Android runtimes.
-   */
-  private interface ByteArrayCopier {
-    /**
-     * Copies the specified range of the specified array into a new array
-     */
-    byte[] copyFrom(byte[] bytes, int offset, int size);
-  }
-
-  /** Implementation of {@code ByteArrayCopier} which uses {@link System#arraycopy}. */
-  private static final class SystemByteArrayCopier implements ByteArrayCopier {
-    @Override
-    public byte[] copyFrom(byte[] bytes, int offset, int size) {
-      byte[] copy = new byte[size];
-      System.arraycopy(bytes, offset, copy, 0, size);
-      return copy;
-    }
-  }
-
-  /** Implementation of {@code ByteArrayCopier} which uses {@link Arrays#copyOfRange}. */
-  private static final class ArraysByteArrayCopier implements ByteArrayCopier {
-    @Override
-    public byte[] copyFrom(byte[] bytes, int offset, int size) {
-      return Arrays.copyOfRange(bytes, offset, offset + size);
-    }
-  }
-
-  private static final ByteArrayCopier byteArrayCopier;
-  static {
-    boolean isAndroid = true;
-    try {
-      Class.forName("android.content.Context");
-    } catch (ClassNotFoundException e) {
-      isAndroid = false;
-    }
-
-    byteArrayCopier = isAndroid ? new SystemByteArrayCopier() : new ArraysByteArrayCopier();
-  }
-
-  /**
-   * Cached hash value. Intentionally accessed via a data race, which
-   * is safe because of the Java Memory Model's "no out-of-thin-air values"
-   * guarantees for ints. A value of 0 implies that the hash has not been set.
-   */
-  private int hash = 0;
-
-  // This constructor is here to prevent subclassing outside of this package,
-  ByteString() {}
-
-  /**
-   * Gets the byte at the given index. This method should be used only for
-   * random access to individual bytes. To access bytes sequentially, use the
-   * {@link ByteIterator} returned by {@link #iterator()}, and call {@link
-   * #substring(int, int)} first if necessary.
-   *
-   * @param index index of byte
-   * @return the value
-   * @throws IndexOutOfBoundsException {@code index < 0 or index >= size}
-   */
-  public abstract byte byteAt(int index);
-
-  /**
-   * Return a {@link ByteString.ByteIterator} over the bytes in the ByteString.
-   * To avoid auto-boxing, you may get the iterator manually and call
-   * {@link ByteIterator#nextByte()}.
-   *
-   * @return the iterator
-   */
-  @Override
-  public final ByteIterator iterator() {
-    return new ByteIterator() {
-      private int position = 0;
-      private final int limit = size();
-
-      @Override
-      public boolean hasNext() {
-        return position < limit;
-      }
-
-      @Override
-      public Byte next() {
-        // Boxing calls Byte.valueOf(byte), which does not instantiate.
-        return nextByte();
-      }
-
-      @Override
-      public byte nextByte() {
-        try {
-          return byteAt(position++);
-        } catch (IndexOutOfBoundsException e) {
-          throw new NoSuchElementException(e.getMessage());
-        }
-      }
-
-      @Override
-      public void remove() {
-        throw new UnsupportedOperationException();
-      }
-    };
-  }
-
-  /**
-   * This interface extends {@code Iterator<Byte>}, so that we can return an
-   * unboxed {@code byte}.
-   */
-  public interface ByteIterator extends Iterator<Byte> {
-    /**
-     * An alternative to {@link Iterator#next()} that returns an
-     * unboxed primitive {@code byte}.
-     *
-     * @return the next {@code byte} in the iteration
-     * @throws NoSuchElementException if the iteration has no more elements
-     */
-    byte nextByte();
-  }
-
-  /**
-   * Gets the number of bytes.
-   *
-   * @return size in bytes
-   */
-  public abstract int size();
-
-  /**
-   * Returns {@code true} if the size is {@code 0}, {@code false} otherwise.
-   *
-   * @return true if this is zero bytes long
-   */
-  public final boolean isEmpty() {
-    return size() == 0;
-  }
-
-  // =================================================================
-  // ByteString -> substring
-
-  /**
-   * Return the substring from {@code beginIndex}, inclusive, to the end of the
-   * string.
-   *
-   * @param beginIndex start at this index
-   * @return substring sharing underlying data
-   * @throws IndexOutOfBoundsException if {@code beginIndex < 0} or
-   *     {@code beginIndex > size()}.
-   */
-  public final ByteString substring(int beginIndex) {
-    return substring(beginIndex, size());
-  }
-
-  /**
-   * Return the substring from {@code beginIndex}, inclusive, to {@code
-   * endIndex}, exclusive.
-   *
-   * @param beginIndex start at this index
-   * @param endIndex   the last character is the one before this index
-   * @return substring sharing underlying data
-   * @throws IndexOutOfBoundsException if {@code beginIndex < 0},
-   *     {@code endIndex > size()}, or {@code beginIndex > endIndex}.
-   */
-  public abstract ByteString substring(int beginIndex, int endIndex);
-
-  /**
-   * Tests if this bytestring starts with the specified prefix.
-   * Similar to {@link String#startsWith(String)}
-   *
-   * @param prefix the prefix.
-   * @return <code>true</code> if the byte sequence represented by the
-   *         argument is a prefix of the byte sequence represented by
-   *         this string; <code>false</code> otherwise.
-   */
-  public final boolean startsWith(ByteString prefix) {
-    return size() >= prefix.size() &&
-           substring(0, prefix.size()).equals(prefix);
-  }
-
-  /**
-   * Tests if this bytestring ends with the specified suffix.
-   * Similar to {@link String#endsWith(String)}
-   *
-   * @param suffix the suffix.
-   * @return <code>true</code> if the byte sequence represented by the
-   *         argument is a suffix of the byte sequence represented by
-   *         this string; <code>false</code> otherwise.
-   */
-  public final boolean endsWith(ByteString suffix) {
-    return size() >= suffix.size() &&
-        substring(size() - suffix.size()).equals(suffix);
-  }
-
-  // =================================================================
-  // byte[] -> ByteString
-
-  /**
-   * Copies the given bytes into a {@code ByteString}.
-   *
-   * @param bytes source array
-   * @param offset offset in source array
-   * @param size number of bytes to copy
-   * @return new {@code ByteString}
-   */
-  public static ByteString copyFrom(byte[] bytes, int offset, int size) {
-    return new LiteralByteString(byteArrayCopier.copyFrom(bytes, offset, size));
-  }
-
-  /**
-   * Copies the given bytes into a {@code ByteString}.
-   *
-   * @param bytes to copy
-   * @return new {@code ByteString}
-   */
-  public static ByteString copyFrom(byte[] bytes) {
-    return copyFrom(bytes, 0, bytes.length);
-  }
-
-  /**
-   * Wraps the given bytes into a {@code ByteString}. Intended for internal only usage.
-   */
-  static ByteString wrap(ByteBuffer buffer) {
-    if (buffer.hasArray()) {
-      final int offset = buffer.arrayOffset();
-      return ByteString.wrap(buffer.array(), offset + buffer.position(), buffer.remaining());
-    } else {
-      return new NioByteString(buffer);
-    }
-  }
-
-  /**
-   * Wraps the given bytes into a {@code ByteString}. Intended for internal only usage.
-   */
-  static ByteString wrap(ByteInput buffer, int offset, int length) {
-    return new ByteInputByteString(buffer, offset, length);
-  }
-
-  /**
-   * Wraps the given bytes into a {@code ByteString}. Intended for internal only
-   * usage to force a classload of ByteString before LiteralByteString.
-   */
-  static ByteString wrap(byte[] bytes) {
-    // TODO(dweis): Return EMPTY when bytes are empty to reduce allocations?
-    return new LiteralByteString(bytes);
-  }
-
-  /**
-   * Wraps the given bytes into a {@code ByteString}. Intended for internal only
-   * usage to force a classload of ByteString before BoundedByteString and
-   * LiteralByteString.
-   */
-  static ByteString wrap(byte[] bytes, int offset, int length) {
-    return new BoundedByteString(bytes, offset, length);
-  }
-
-  /**
-   * Copies the next {@code size} bytes from a {@code java.nio.ByteBuffer} into
-   * a {@code ByteString}.
-   *
-   * @param bytes source buffer
-   * @param size number of bytes to copy
-   * @return new {@code ByteString}
-   */
-  public static ByteString copyFrom(ByteBuffer bytes, int size) {
-    byte[] copy = new byte[size];
-    bytes.get(copy);
-    return new LiteralByteString(copy);
-  }
-
-  /**
-   * Copies the remaining bytes from a {@code java.nio.ByteBuffer} into
-   * a {@code ByteString}.
-   *
-   * @param bytes sourceBuffer
-   * @return new {@code ByteString}
-   */
-  public static ByteString copyFrom(ByteBuffer bytes) {
-    return copyFrom(bytes, bytes.remaining());
-  }
-
-  /**
-   * Encodes {@code text} into a sequence of bytes using the named charset
-   * and returns the result as a {@code ByteString}.
-   *
-   * @param text source string
-   * @param charsetName encoding to use
-   * @return new {@code ByteString}
-   * @throws UnsupportedEncodingException if the encoding isn't found
-   */
-  public static ByteString copyFrom(String text, String charsetName)
-      throws UnsupportedEncodingException {
-    return new LiteralByteString(text.getBytes(charsetName));
-  }
-
-  /**
-   * Encodes {@code text} into a sequence of bytes using the named charset
-   * and returns the result as a {@code ByteString}.
-   *
-   * @param text source string
-   * @param charset encode using this charset
-   * @return new {@code ByteString}
-   */
-  public static ByteString copyFrom(String text, Charset charset) {
-    return new LiteralByteString(text.getBytes(charset));
-  }
-
-  /**
-   * Encodes {@code text} into a sequence of UTF-8 bytes and returns the
-   * result as a {@code ByteString}.
-   *
-   * @param text source string
-   * @return new {@code ByteString}
-   */
-  public static ByteString copyFromUtf8(String text) {
-    return new LiteralByteString(text.getBytes(Internal.UTF_8));
-  }
-
-  // =================================================================
-  // InputStream -> ByteString
-
-  /**
-   * Completely reads the given stream's bytes into a
-   * {@code ByteString}, blocking if necessary until all bytes are
-   * read through to the end of the stream.
-   *
-   * <b>Performance notes:</b> The returned {@code ByteString} is an
-   * immutable tree of byte arrays ("chunks") of the stream data.  The
-   * first chunk is small, with subsequent chunks each being double
-   * the size, up to 8K.
-   *
-   * <p>Each byte read from the input stream will be copied twice to ensure
-   * that the resulting ByteString is truly immutable.
-   *
-   * @param streamToDrain The source stream, which is read completely
-   *     but not closed.
-   * @return A new {@code ByteString} which is made up of chunks of
-   *     various sizes, depending on the behavior of the underlying
-   *     stream.
-   * @throws IOException IOException is thrown if there is a problem
-   *     reading the underlying stream.
-   */
-  public static ByteString readFrom(InputStream streamToDrain)
-      throws IOException {
-    return readFrom(streamToDrain, MIN_READ_FROM_CHUNK_SIZE, MAX_READ_FROM_CHUNK_SIZE);
-  }
-
-  /**
-   * Completely reads the given stream's bytes into a
-   * {@code ByteString}, blocking if necessary until all bytes are
-   * read through to the end of the stream.
-   *
-   * <b>Performance notes:</b> The returned {@code ByteString} is an
-   * immutable tree of byte arrays ("chunks") of the stream data.  The
-   * chunkSize parameter sets the size of these byte arrays.
-   *
-   * <p>Each byte read from the input stream will be copied twice to ensure
-   * that the resulting ByteString is truly immutable.
-   *
-   * @param streamToDrain The source stream, which is read completely
-   *     but not closed.
-   * @param chunkSize The size of the chunks in which to read the
-   *     stream.
-   * @return A new {@code ByteString} which is made up of chunks of
-   *     the given size.
-   * @throws IOException IOException is thrown if there is a problem
-   *     reading the underlying stream.
-   */
-  public static ByteString readFrom(InputStream streamToDrain, int chunkSize)
-      throws IOException {
-    return readFrom(streamToDrain, chunkSize, chunkSize);
-  }
-
-  // Helper method that takes the chunk size range as a parameter.
-  public static ByteString readFrom(InputStream streamToDrain, int minChunkSize,
-      int maxChunkSize) throws IOException {
-    Collection<ByteString> results = new ArrayList<ByteString>();
-
-    // copy the inbound bytes into a list of chunks; the chunk size
-    // grows exponentially to support both short and long streams.
-    int chunkSize = minChunkSize;
-    while (true) {
-      ByteString chunk = readChunk(streamToDrain, chunkSize);
-      if (chunk == null) {
-        break;
-      }
-      results.add(chunk);
-      chunkSize = Math.min(chunkSize * 2, maxChunkSize);
-    }
-
-    return ByteString.copyFrom(results);
-  }
-
-  /**
-   * Blocks until a chunk of the given size can be made from the
-   * stream, or EOF is reached.  Calls read() repeatedly in case the
-   * given stream implementation doesn't completely fill the given
-   * buffer in one read() call.
-   *
-   * @return A chunk of the desired size, or else a chunk as large as
-   * was available when end of stream was reached. Returns null if the
-   * given stream had no more data in it.
-   */
-  private static ByteString readChunk(InputStream in, final int chunkSize)
-      throws IOException {
-      final byte[] buf = new byte[chunkSize];
-      int bytesRead = 0;
-      while (bytesRead < chunkSize) {
-        final int count = in.read(buf, bytesRead, chunkSize - bytesRead);
-        if (count == -1) {
-          break;
-        }
-        bytesRead += count;
-      }
-
-      if (bytesRead == 0) {
-        return null;
-      }
-
-      // Always make a copy since InputStream could steal a reference to buf.
-      return ByteString.copyFrom(buf, 0, bytesRead);
-  }
-
-  // =================================================================
-  // Multiple ByteStrings -> One ByteString
-
-  /**
-   * Concatenate the given {@code ByteString} to this one. Short concatenations,
-   * of total size smaller than {@link ByteString#CONCATENATE_BY_COPY_SIZE}, are
-   * produced by copying the underlying bytes (as per Rope.java, <a
-   * href="http://www.cs.ubc.ca/local/reading/proceedings/spe91-95/spe/vol25/issue12/spe986.pdf">
-   * BAP95 </a>. In general, the concatenate involves no copying.
-   *
-   * @param other string to concatenate
-   * @return a new {@code ByteString} instance
-   */
-  public final ByteString concat(ByteString other) {
-    if (Integer.MAX_VALUE - size() < other.size()) {
-      throw new IllegalArgumentException("ByteString would be too long: " +
-          size() + "+" + other.size());
-    }
-
-    return RopeByteString.concatenate(this, other);
-  }
-
-  /**
-   * Concatenates all byte strings in the iterable and returns the result.
-   * This is designed to run in O(list size), not O(total bytes).
-   *
-   * <p>The returned {@code ByteString} is not necessarily a unique object.
-   * If the list is empty, the returned object is the singleton empty
-   * {@code ByteString}.  If the list has only one element, that
-   * {@code ByteString} will be returned without copying.
-   *
-   * @param byteStrings strings to be concatenated
-   * @return new {@code ByteString}
-   */
-  public static ByteString copyFrom(Iterable<ByteString> byteStrings) {
-    // Determine the size;
-    final int size;
-    if (!(byteStrings instanceof Collection)) {
-      int tempSize = 0;
-      for (Iterator<ByteString> iter = byteStrings.iterator(); iter.hasNext();
-          iter.next(), ++tempSize) {
-      }
-      size = tempSize;
-    } else {
-      size = ((Collection<ByteString>) byteStrings).size();
-    }
-
-    if (size == 0) {
-      return EMPTY;
-    }
-
-    return balancedConcat(byteStrings.iterator(), size);
-  }
-
-  // Internal function used by copyFrom(Iterable<ByteString>).
-  // Create a balanced concatenation of the next "length" elements from the
-  // iterable.
-  private static ByteString balancedConcat(Iterator<ByteString> iterator, int length) {
-    if (length < 1) {
-      throw new IllegalArgumentException(String.format("length (%s) must be >= 1", length));
-    }
-    ByteString result;
-    if (length == 1) {
-      result = iterator.next();
-    } else {
-      int halfLength = length >>> 1;
-      ByteString left = balancedConcat(iterator, halfLength);
-      ByteString right = balancedConcat(iterator, length - halfLength);
-      result = left.concat(right);
-    }
-    return result;
-  }
-
-  // =================================================================
-  // ByteString -> byte[]
-
-  /**
-   * Copies bytes into a buffer at the given offset.
-   *
-   * @param target buffer to copy into
-   * @param offset in the target buffer
-   * @throws IndexOutOfBoundsException if the offset is negative or too large
-   */
-  public void copyTo(byte[] target, int offset) {
-    copyTo(target, 0, offset, size());
-  }
-
-  /**
-   * Copies bytes into a buffer.
-   *
-   * @param target       buffer to copy into
-   * @param sourceOffset offset within these bytes
-   * @param targetOffset offset within the target buffer
-   * @param numberToCopy number of bytes to copy
-   * @throws IndexOutOfBoundsException if an offset or size is negative or too
-   *     large
-   */
-  public final void copyTo(byte[] target, int sourceOffset, int targetOffset,
-      int numberToCopy) {
-    checkRange(sourceOffset, sourceOffset + numberToCopy, size());
-    checkRange(targetOffset, targetOffset + numberToCopy, target.length);
-    if (numberToCopy > 0) {
-      copyToInternal(target, sourceOffset, targetOffset, numberToCopy);
-    }
-  }
-
-  /**
-   * Internal (package private) implementation of
-   * {@link #copyTo(byte[],int,int,int)}.
-   * It assumes that all error checking has already been performed and that
-   * {@code numberToCopy > 0}.
-   */
-  protected abstract void copyToInternal(byte[] target, int sourceOffset,
-      int targetOffset, int numberToCopy);
-
-  /**
-   * Copies bytes into a ByteBuffer.
-   *
-   * @param target ByteBuffer to copy into.
-   * @throws java.nio.ReadOnlyBufferException if the {@code target} is read-only
-   * @throws java.nio.BufferOverflowException if the {@code target}'s
-   *     remaining() space is not large enough to hold the data.
-   */
-  public abstract void copyTo(ByteBuffer target);
-
-  /**
-   * Copies bytes to a {@code byte[]}.
-   *
-   * @return copied bytes
-   */
-  public final byte[] toByteArray() {
-    final int size = size();
-    if (size == 0) {
-      return Internal.EMPTY_BYTE_ARRAY;
-    }
-    byte[] result = new byte[size];
-    copyToInternal(result, 0, 0, size);
-    return result;
-  }
-
-  /**
-   * Writes a copy of the contents of this byte string to the specified output stream argument.
-   *
-   * @param  out  the output stream to which to write the data.
-   * @throws IOException  if an I/O error occurs.
-   */
-  public abstract void writeTo(OutputStream out) throws IOException;
-
-  /**
-   * Writes a specified part of this byte string to an output stream.
-   *
-   * @param  out  the output stream to which to write the data.
-   * @param  sourceOffset offset within these bytes
-   * @param  numberToWrite number of bytes to write
-   * @throws IOException  if an I/O error occurs.
-   * @throws IndexOutOfBoundsException if an offset or size is negative or too large
-   */
-  final void writeTo(OutputStream out, int sourceOffset, int numberToWrite)
-      throws IOException {
-    checkRange(sourceOffset, sourceOffset + numberToWrite, size());
-    if (numberToWrite > 0) {
-      writeToInternal(out, sourceOffset, numberToWrite);
-    }
-  }
-
-  /**
-   * Internal version of {@link #writeTo(OutputStream,int,int)} that assumes
-   * all error checking has already been done.
-   */
-  abstract void writeToInternal(OutputStream out, int sourceOffset, int numberToWrite)
-      throws IOException;
-
-  /**
-   * Writes this {@link ByteString} to the provided {@link ByteOutput}. Calling
-   * this method may result in multiple operations on the target {@link ByteOutput}.
-   *
-   * <p>This method may expose internal backing buffers of the {@link ByteString} to the {@link
-   * ByteOutput} in order to avoid additional copying overhead. It would be possible for a malicious
-   * {@link ByteOutput} to corrupt the {@link ByteString}. Use with caution!
-   *
-   * @param  byteOutput  the output target to receive the bytes
-   * @throws IOException  if an I/O error occurs
-   * @see UnsafeByteOperations#unsafeWriteTo(ByteString, ByteOutput)
-   */
-  abstract void writeTo(ByteOutput byteOutput) throws IOException;
-
-
-  /**
-   * Constructs a read-only {@code java.nio.ByteBuffer} whose content
-   * is equal to the contents of this byte string.
-   * The result uses the same backing array as the byte string, if possible.
-   *
-   * @return wrapped bytes
-   */
-  public abstract ByteBuffer asReadOnlyByteBuffer();
-
-  /**
-   * Constructs a list of read-only {@code java.nio.ByteBuffer} objects
-   * such that the concatenation of their contents is equal to the contents
-   * of this byte string.  The result uses the same backing arrays as the
-   * byte string.
-   * <p>
-   * By returning a list, implementations of this method may be able to avoid
-   * copying even when there are multiple backing arrays.
-   *
-   * @return a list of wrapped bytes
-   */
-  public abstract List<ByteBuffer> asReadOnlyByteBufferList();
-
-  /**
-   * Constructs a new {@code String} by decoding the bytes using the
-   * specified charset.
-   *
-   * @param charsetName encode using this charset
-   * @return new string
-   * @throws UnsupportedEncodingException if charset isn't recognized
-   */
-  public final String toString(String charsetName)
-      throws UnsupportedEncodingException {
-    try {
-      return toString(Charset.forName(charsetName));
-    } catch (UnsupportedCharsetException e) {
-      UnsupportedEncodingException exception = new UnsupportedEncodingException(charsetName);
-      exception.initCause(e);
-      throw exception;
-    }
-  }
-
-  /**
-   * Constructs a new {@code String} by decoding the bytes using the
-   * specified charset. Returns the same empty String if empty.
-   *
-   * @param charset encode using this charset
-   * @return new string
-   */
-  public final String toString(Charset charset) {
-    return size() == 0 ? "" : toStringInternal(charset);
-  }
-
-  /**
-   * Constructs a new {@code String} by decoding the bytes using the
-   * specified charset.
-   *
-   * @param charset encode using this charset
-   * @return new string
-   */
-  protected abstract String toStringInternal(Charset charset);
-
-  // =================================================================
-  // UTF-8 decoding
-
-  /**
-   * Constructs a new {@code String} by decoding the bytes as UTF-8.
-   *
-   * @return new string using UTF-8 encoding
-   */
-  public final String toStringUtf8() {
-    return toString(Internal.UTF_8);
-  }
-
-  /**
-   * Tells whether this {@code ByteString} represents a well-formed UTF-8
-   * byte sequence, such that the original bytes can be converted to a
-   * String object and then round tripped back to bytes without loss.
-   *
-   * <p>More precisely, returns {@code true} whenever: <pre> {@code
-   * Arrays.equals(byteString.toByteArray(),
-   *     new String(byteString.toByteArray(), "UTF-8").getBytes("UTF-8"))
-   * }</pre>
-   *
-   * <p>This method returns {@code false} for "overlong" byte sequences,
-   * as well as for 3-byte sequences that would map to a surrogate
-   * character, in accordance with the restricted definition of UTF-8
-   * introduced in Unicode 3.1.  Note that the UTF-8 decoder included in
-   * Oracle's JDK has been modified to also reject "overlong" byte
-   * sequences, but (as of 2011) still accepts 3-byte surrogate
-   * character byte sequences.
-   *
-   * <p>See the Unicode Standard,<br>
-   * Table 3-6. <em>UTF-8 Bit Distribution</em>,<br>
-   * Table 3-7. <em>Well Formed UTF-8 Byte Sequences</em>.
-   *
-   * @return whether the bytes in this {@code ByteString} are a
-   * well-formed UTF-8 byte sequence
-   */
-  public abstract boolean isValidUtf8();
-
-  /**
-   * Tells whether the given byte sequence is a well-formed, malformed, or
-   * incomplete UTF-8 byte sequence.  This method accepts and returns a partial
-   * state result, allowing the bytes for a complete UTF-8 byte sequence to be
-   * composed from multiple {@code ByteString} segments.
-   *
-   * @param state either {@code 0} (if this is the initial decoding operation)
-   *     or the value returned from a call to a partial decoding method for the
-   *     previous bytes
-   * @param offset offset of the first byte to check
-   * @param length number of bytes to check
-   *
-   * @return {@code -1} if the partial byte sequence is definitely malformed,
-   * {@code 0} if it is well-formed (no additional input needed), or, if the
-   * byte sequence is "incomplete", i.e. apparently terminated in the middle of
-   * a character, an opaque integer "state" value containing enough information
-   * to decode the character when passed to a subsequent invocation of a
-   * partial decoding method.
-   */
-  protected abstract int partialIsValidUtf8(int state, int offset, int length);
-
-  // =================================================================
-  // equals() and hashCode()
-
-  @Override
-  public abstract boolean equals(Object o);
-
-  /**
-   * Base class for leaf {@link ByteString}s (i.e. non-ropes).
-   */
-  abstract static class LeafByteString extends ByteString {
-    @Override
-    protected final int getTreeDepth() {
-      return 0;
-    }
-
-    @Override
-    protected final boolean isBalanced() {
-      return true;
-    }
-
-
-    /**
-     * Check equality of the substring of given length of this object starting at
-     * zero with another {@code ByteString} substring starting at offset.
-     *
-     * @param other  what to compare a substring in
-     * @param offset offset into other
-     * @param length number of bytes to compare
-     * @return true for equality of substrings, else false.
-     */
-    abstract boolean equalsRange(ByteString other, int offset, int length);
-  }
-
-  /**
-   * Compute the hashCode using the traditional algorithm from {@link
-   * ByteString}.
-   *
-   * @return hashCode value
-   */
-  @Override
-  public final int hashCode() {
-    int h = hash;
-
-    if (h == 0) {
-      int size = size();
-      h = partialHash(size, 0, size);
-      if (h == 0) {
-        h = 1;
-      }
-      hash = h;
-    }
-    return h;
-  }
-
-  // =================================================================
-  // Input stream
-
-  /**
-   * Creates an {@code InputStream} which can be used to read the bytes.
-   * <p>
-   * The {@link InputStream} returned by this method is guaranteed to be
-   * completely non-blocking.  The method {@link InputStream#available()}
-   * returns the number of bytes remaining in the stream. The methods
-   * {@link InputStream#read(byte[])}, {@link InputStream#read(byte[],int,int)}
-   * and {@link InputStream#skip(long)} will read/skip as many bytes as are
-   * available.  The method {@link InputStream#markSupported()} returns
-   * {@code true}.
-   * <p>
-   * The methods in the returned {@link InputStream} might <b>not</b> be
-   * thread safe.
-   *
-   * @return an input stream that returns the bytes of this byte string.
-   */
-  public abstract InputStream newInput();
-
-  /**
-   * Creates a {@link CodedInputStream} which can be used to read the bytes.
-   * Using this is often more efficient than creating a {@link CodedInputStream}
-   * that wraps the result of {@link #newInput()}.
-   *
-   * @return stream based on wrapped data
-   */
-  public abstract CodedInputStream newCodedInput();
-
-  // =================================================================
-  // Output stream
-
-  /**
-   * Creates a new {@link Output} with the given initial capacity. Call {@link
-   * Output#toByteString()} to create the {@code ByteString} instance.
-   * <p>
-   * A {@link ByteString.Output} offers the same functionality as a
-   * {@link ByteArrayOutputStream}, except that it returns a {@link ByteString}
-   * rather than a {@code byte} array.
-   *
-   * @param initialCapacity estimate of number of bytes to be written
-   * @return {@code OutputStream} for building a {@code ByteString}
-   */
-  public static Output newOutput(int initialCapacity) {
-    return new Output(initialCapacity);
-  }
-
-  /**
-   * Creates a new {@link Output}. Call {@link Output#toByteString()} to create
-   * the {@code ByteString} instance.
-   * <p>
-   * A {@link ByteString.Output} offers the same functionality as a
-   * {@link ByteArrayOutputStream}, except that it returns a {@link ByteString}
-   * rather than a {@code byte array}.
-   *
-   * @return {@code OutputStream} for building a {@code ByteString}
-   */
-  public static Output newOutput() {
-    return new Output(CONCATENATE_BY_COPY_SIZE);
-  }
-
-  /**
-   * Outputs to a {@code ByteString} instance. Call {@link #toByteString()} to
-   * create the {@code ByteString} instance.
-   */
-  public static final class Output extends OutputStream {
-    // Implementation note.
-    // The public methods of this class must be synchronized.  ByteStrings
-    // are guaranteed to be immutable.  Without some sort of locking, it could
-    // be possible for one thread to call toByteSring(), while another thread
-    // is still modifying the underlying byte array.
-
-    private static final byte[] EMPTY_BYTE_ARRAY = new byte[0];
-    // argument passed by user, indicating initial capacity.
-    private final int initialCapacity;
-    // ByteStrings to be concatenated to create the result
-    private final ArrayList<ByteString> flushedBuffers;
-    // Total number of bytes in the ByteStrings of flushedBuffers
-    private int flushedBuffersTotalBytes;
-    // Current buffer to which we are writing
-    private byte[] buffer;
-    // Location in buffer[] to which we write the next byte.
-    private int bufferPos;
-
-    /**
-     * Creates a new ByteString output stream with the specified
-     * initial capacity.
-     *
-     * @param initialCapacity  the initial capacity of the output stream.
-     */
-    Output(int initialCapacity) {
-      if (initialCapacity < 0) {
-        throw new IllegalArgumentException("Buffer size < 0");
-      }
-      this.initialCapacity = initialCapacity;
-      this.flushedBuffers = new ArrayList<ByteString>();
-      this.buffer = new byte[initialCapacity];
-    }
-
-    @Override
-    public synchronized void write(int b) {
-      if (bufferPos == buffer.length) {
-        flushFullBuffer(1);
-      }
-      buffer[bufferPos++] = (byte)b;
-    }
-
-    @Override
-    public synchronized void write(byte[] b, int offset, int length)  {
-      if (length <= buffer.length - bufferPos) {
-        // The bytes can fit into the current buffer.
-        System.arraycopy(b, offset, buffer, bufferPos, length);
-        bufferPos += length;
-      } else {
-        // Use up the current buffer
-        int copySize  = buffer.length - bufferPos;
-        System.arraycopy(b, offset, buffer, bufferPos, copySize);
-        offset += copySize;
-        length -= copySize;
-        // Flush the buffer, and get a new buffer at least big enough to cover
-        // what we still need to output
-        flushFullBuffer(length);
-        System.arraycopy(b, offset, buffer, 0 /* count */, length);
-        bufferPos = length;
-      }
-    }
-
-    /**
-     * Creates a byte string. Its size is the current size of this output
-     * stream and its output has been copied to it.
-     *
-     * @return  the current contents of this output stream, as a byte string.
-     */
-    public synchronized ByteString toByteString() {
-      flushLastBuffer();
-      return ByteString.copyFrom(flushedBuffers);
-    }
-
-    /**
-     * Implement java.util.Arrays.copyOf() for jdk 1.5.
-     */
-    private byte[] copyArray(byte[] buffer, int length) {
-      byte[] result = new byte[length];
-      System.arraycopy(buffer, 0, result, 0, Math.min(buffer.length, length));
-      return result;
-    }
-
-    /**
-     * Writes the complete contents of this byte array output stream to
-     * the specified output stream argument.
-     *
-     * @param out the output stream to which to write the data.
-     * @throws IOException  if an I/O error occurs.
-     */
-    public void writeTo(OutputStream out) throws IOException {
-      ByteString[] cachedFlushBuffers;
-      byte[] cachedBuffer;
-      int cachedBufferPos;
-      synchronized (this) {
-        // Copy the information we need into local variables so as to hold
-        // the lock for as short a time as possible.
-        cachedFlushBuffers =
-            flushedBuffers.toArray(new ByteString[flushedBuffers.size()]);
-        cachedBuffer = buffer;
-        cachedBufferPos = bufferPos;
-      }
-      for (ByteString byteString : cachedFlushBuffers) {
-        byteString.writeTo(out);
-      }
-
-      out.write(copyArray(cachedBuffer, cachedBufferPos));
-    }
-
-    /**
-     * Returns the current size of the output stream.
-     *
-     * @return  the current size of the output stream
-     */
-    public synchronized int size() {
-      return flushedBuffersTotalBytes + bufferPos;
-    }
-
-    /**
-     * Resets this stream, so that all currently accumulated output in the
-     * output stream is discarded. The output stream can be used again,
-     * reusing the already allocated buffer space.
-     */
-    public synchronized void reset() {
-      flushedBuffers.clear();
-      flushedBuffersTotalBytes = 0;
-      bufferPos = 0;
-    }
-
-    @Override
-    public String toString() {
-      return String.format("<ByteString.Output@%s size=%d>",
-          Integer.toHexString(System.identityHashCode(this)), size());
-    }
-
-    /**
-     * Internal function used by writers.  The current buffer is full, and the
-     * writer needs a new buffer whose size is at least the specified minimum
-     * size.
-     */
-    private void flushFullBuffer(int minSize)  {
-      flushedBuffers.add(new LiteralByteString(buffer));
-      flushedBuffersTotalBytes += buffer.length;
-      // We want to increase our total capacity by 50%, but as a minimum,
-      // the new buffer should also at least be >= minSize and
-      // >= initial Capacity.
-      int newSize = Math.max(initialCapacity,
-          Math.max(minSize, flushedBuffersTotalBytes >>> 1));
-      buffer = new byte[newSize];
-      bufferPos = 0;
-    }
-
-    /**
-     * Internal function used by {@link #toByteString()}. The current buffer may
-     * or may not be full, but it needs to be flushed.
-     */
-    private void flushLastBuffer()  {
-      if (bufferPos < buffer.length) {
-        if (bufferPos > 0) {
-          byte[] bufferCopy = copyArray(buffer, bufferPos);
-          flushedBuffers.add(new LiteralByteString(bufferCopy));
-        }
-        // We reuse this buffer for further writes.
-      } else {
-        // Buffer is completely full.  Huzzah.
-        flushedBuffers.add(new LiteralByteString(buffer));
-        // 99% of the time, we're not going to use this OutputStream again.
-        // We set buffer to an empty byte stream so that we're handling this
-        // case without wasting space.  In the rare case that more writes
-        // *do* occur, this empty buffer will be flushed and an appropriately
-        // sized new buffer will be created.
-        buffer = EMPTY_BYTE_ARRAY;
-      }
-      flushedBuffersTotalBytes += bufferPos;
-      bufferPos = 0;
-    }
-  }
-
-  /**
-   * Constructs a new {@code ByteString} builder, which allows you to
-   * efficiently construct a {@code ByteString} by writing to a {@link
-   * CodedOutputStream}. Using this is much more efficient than calling {@code
-   * newOutput()} and wrapping that in a {@code CodedOutputStream}.
-   *
-   * <p>This is package-private because it's a somewhat confusing interface.
-   * Users can call {@link Message#toByteString()} instead of calling this
-   * directly.
-   *
-   * @param size The target byte size of the {@code ByteString}.  You must write
-   *     exactly this many bytes before building the result.
-   * @return the builder
-   */
-  static CodedBuilder newCodedBuilder(int size) {
-    return new CodedBuilder(size);
-  }
-
-  /** See {@link ByteString#newCodedBuilder(int)}. */
-  static final class CodedBuilder {
-    private final CodedOutputStream output;
-    private final byte[] buffer;
-
-    private CodedBuilder(int size) {
-      buffer = new byte[size];
-      output = CodedOutputStream.newInstance(buffer);
-    }
-
-    public ByteString build() {
-      output.checkNoSpaceLeft();
-
-      // We can be confident that the CodedOutputStream will not modify the
-      // underlying bytes anymore because it already wrote all of them.  So,
-      // no need to make a copy.
-      return new LiteralByteString(buffer);
-    }
-
-    public CodedOutputStream getCodedOutput() {
-      return output;
-    }
-  }
-
-  // =================================================================
-  // Methods {@link RopeByteString} needs on instances, which aren't part of the
-  // public API.
-
-  /**
-   * Return the depth of the tree representing this {@code ByteString}, if any,
-   * whose root is this node. If this is a leaf node, return 0.
-   *
-   * @return tree depth or zero
-   */
-  protected abstract int getTreeDepth();
-
-  /**
-   * Return {@code true} if this ByteString is literal (a leaf node) or a
-   * flat-enough tree in the sense of {@link RopeByteString}.
-   *
-   * @return true if the tree is flat enough
-   */
-  protected abstract boolean isBalanced();
-
-  /**
-   * Return the cached hash code if available.
-   *
-   * @return value of cached hash code or 0 if not computed yet
-   */
-  protected final int peekCachedHashCode() {
-    return hash;
-  }
-
-  /**
-   * Compute the hash across the value bytes starting with the given hash, and
-   * return the result.  This is used to compute the hash across strings
-   * represented as a set of pieces by allowing the hash computation to be
-   * continued from piece to piece.
-   *
-   * @param h starting hash value
-   * @param offset offset into this value to start looking at data values
-   * @param length number of data values to include in the hash computation
-   * @return ending hash value
-   */
-  protected abstract int partialHash(int h, int offset, int length);
-
-  /**
-   * Checks that the given index falls within the specified array size.
-   *
-   * @param index the index position to be tested
-   * @param size the length of the array
-   * @throws IndexOutOfBoundsException if the index does not fall within the array.
-   */
-  static void checkIndex(int index, int size) {
-    if ((index | (size - (index + 1))) < 0) {
-      if (index < 0) {
-        throw new ArrayIndexOutOfBoundsException("Index < 0: " + index);
-      }
-      throw new ArrayIndexOutOfBoundsException("Index > length: " + index + ", " + size);
-    }
-  }
-
-  /**
-   * Checks that the given range falls within the bounds of an array
-   *
-   * @param startIndex the start index of the range (inclusive)
-   * @param endIndex the end index of the range (exclusive)
-   * @param size the size of the array.
-   * @return the length of the range.
-   * @throws IndexOutOfBoundsException some or all of the range falls outside of the array.
-   */
-  static int checkRange(int startIndex, int endIndex, int size) {
-    final int length = endIndex - startIndex;
-    if ((startIndex | endIndex | length | (size - endIndex)) < 0) {
-      if (startIndex < 0) {
-        throw new IndexOutOfBoundsException("Beginning index: " + startIndex + " < 0");
-      }
-      if (endIndex < startIndex) {
-        throw new IndexOutOfBoundsException(
-            "Beginning index larger than ending index: " + startIndex + ", " + endIndex);
-      }
-      // endIndex >= size
-      throw new IndexOutOfBoundsException("End index: " + endIndex + " >= " + size);
-    }
-    return length;
-  }
-
-  @Override
-  public final String toString() {
-    return String.format("<ByteString@%s size=%d>",
-        Integer.toHexString(System.identityHashCode(this)), size());
-  }
-
-  /**
-   * This class implements a {@link org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString} backed by a
-   * single array of bytes, contiguous in memory. It supports substring by
-   * pointing to only a sub-range of the underlying byte array, meaning that a
-   * substring will reference the full byte-array of the string it's made from,
-   * exactly as with {@link String}.
-   *
-   * @author carlanton@google.com (Carl Haverl)
-   */
-  // Keep this class private to avoid deadlocks in classloading across threads as ByteString's
-  // static initializer loads LiteralByteString and another thread loads LiteralByteString.
-  private static class LiteralByteString extends ByteString.LeafByteString {
-    private static final long serialVersionUID = 1L;
-
-    protected final byte[] bytes;
-
-    /**
-     * Creates a {@code LiteralByteString} backed by the given array, without
-     * copying.
-     *
-     * @param bytes array to wrap
-     */
-    LiteralByteString(byte[] bytes) {
-      this.bytes = bytes;
-    }
-
-    @Override
-    public byte byteAt(int index) {
-      // Unlike most methods in this class, this one is a direct implementation
-      // ignoring the potential offset because we need to do range-checking in the
-      // substring case anyway.
-      return bytes[index];
-    }
-
-    @Override
-    public int size() {
-      return bytes.length;
-    }
-
-    // =================================================================
-    // ByteString -> substring
-
-    @Override
-    public final ByteString substring(int beginIndex, int endIndex) {
-      final int length = checkRange(beginIndex, endIndex, size());
-
-      if (length == 0) {
-        return ByteString.EMPTY;
-      }
-
-      return new BoundedByteString(bytes, getOffsetIntoBytes() + beginIndex, length);
-    }
-
-    // =================================================================
-    // ByteString -> byte[]
-
-    @Override
-    protected void copyToInternal(
-        byte[] target, int sourceOffset, int targetOffset, int numberToCopy) {
-      // Optimized form, not for subclasses, since we don't call
-      // getOffsetIntoBytes() or check the 'numberToCopy' parameter.
-      // TODO(nathanmittler): Is not calling getOffsetIntoBytes really saving that much?
-      System.arraycopy(bytes, sourceOffset, target, targetOffset, numberToCopy);
-    }
-
-    @Override
-    public final void copyTo(ByteBuffer target) {
-      target.put(bytes, getOffsetIntoBytes(), size()); // Copies bytes
-    }
-
-    @Override
-    public final ByteBuffer asReadOnlyByteBuffer() {
-      return ByteBuffer.wrap(bytes, getOffsetIntoBytes(), size()).asReadOnlyBuffer();
-    }
-
-    @Override
-    public final List<ByteBuffer> asReadOnlyByteBufferList() {
-      return Collections.singletonList(asReadOnlyByteBuffer());
-    }
-
-    @Override
-    public final void writeTo(OutputStream outputStream) throws IOException {
-      outputStream.write(toByteArray());
-    }
-
-    @Override
-    final void writeToInternal(OutputStream outputStream, int sourceOffset, int numberToWrite)
-        throws IOException {
-      outputStream.write(bytes, getOffsetIntoBytes() + sourceOffset, numberToWrite);
-    }
-
-    @Override
-    final void writeTo(ByteOutput output) throws IOException {
-      output.writeLazy(bytes, getOffsetIntoBytes(), size());
-    }
-
-    @Override
-    protected final String toStringInternal(Charset charset) {
-      return new String(bytes, getOffsetIntoBytes(), size(), charset);
-    }
-
-    // =================================================================
-    // UTF-8 decoding
-
-    @Override
-    public final boolean isValidUtf8() {
-      int offset = getOffsetIntoBytes();
-      return Utf8.isValidUtf8(bytes, offset, offset + size());
-    }
-
-    @Override
-    protected final int partialIsValidUtf8(int state, int offset, int length) {
-      int index = getOffsetIntoBytes() + offset;
-      return Utf8.partialIsValidUtf8(state, bytes, index, index + length);
-    }
-
-    // =================================================================
-    // equals() and hashCode()
-
-    @Override
-    public final boolean equals(Object other) {
-      if (other == this) {
-        return true;
-      }
-      if (!(other instanceof ByteString)) {
-        return false;
-      }
-
-      if (size() != ((ByteString) other).size()) {
-        return false;
-      }
-      if (size() == 0) {
-        return true;
-      }
-
-      if (other instanceof LiteralByteString) {
-        LiteralByteString otherAsLiteral = (LiteralByteString) other;
-        // If we know the hash codes and they are not equal, we know the byte
-        // strings are not equal.
-        int thisHash = peekCachedHashCode();
-        int thatHash = otherAsLiteral.peekCachedHashCode();
-        if (thisHash != 0 && thatHash != 0 && thisHash != thatHash) {
-          return false;
-        }
-
-        return equalsRange((LiteralByteString) other, 0, size());
-      } else {
-        // RopeByteString and NioByteString.
-        return other.equals(this);
-      }
-    }
-
-    /**
-     * Check equality of the substring of given length of this object starting at
-     * zero with another {@code LiteralByteString} substring starting at offset.
-     *
-     * @param other  what to compare a substring in
-     * @param offset offset into other
-     * @param length number of bytes to compare
-     * @return true for equality of substrings, else false.
-     */
-    @Override
-    final boolean equalsRange(ByteString other, int offset, int length) {
-      if (length > other.size()) {
-        throw new IllegalArgumentException("Length too large: " + length + size());
-      }
-      if (offset + length > other.size()) {
-        throw new IllegalArgumentException(
-            "Ran off end of other: " + offset + ", " + length + ", " + other.size());
-      }
-
-      if (other instanceof LiteralByteString) {
-        LiteralByteString lbsOther = (LiteralByteString) other;
-        byte[] thisBytes = bytes;
-        byte[] otherBytes = lbsOther.bytes;
-        int thisLimit = getOffsetIntoBytes() + length;
-        for (
-            int thisIndex = getOffsetIntoBytes(),
-                otherIndex = lbsOther.getOffsetIntoBytes() + offset;
-            (thisIndex < thisLimit); ++thisIndex, ++otherIndex) {
-          if (thisBytes[thisIndex] != otherBytes[otherIndex]) {
-            return false;
-          }
-        }
-        return true;
-      }
-
-      return other.substring(offset, offset + length).equals(substring(0, length));
-    }
-
-    @Override
-    protected final int partialHash(int h, int offset, int length) {
-      return Internal.partialHash(h, bytes, getOffsetIntoBytes() + offset, length);
-    }
-
-    // =================================================================
-    // Input stream
-
-    @Override
-    public final InputStream newInput() {
-      return new ByteArrayInputStream(bytes, getOffsetIntoBytes(), size()); // No copy
-    }
-
-    @Override
-    public final CodedInputStream newCodedInput() {
-      // We trust CodedInputStream not to modify the bytes, or to give anyone
-      // else access to them.
-      return CodedInputStream.newInstance(
-          bytes, getOffsetIntoBytes(), size(), true /* bufferIsImmutable */);
-    }
-
-    // =================================================================
-    // Internal methods
-
-    /**
-     * Offset into {@code bytes[]} to use, non-zero for substrings.
-     *
-     * @return always 0 for this class
-     */
-    protected int getOffsetIntoBytes() {
-      return 0;
-    }
-  }
-
-  /**
-   * This class is used to represent the substring of a {@link ByteString} over a
-   * single byte array. In terms of the public API of {@link ByteString}, you end
-   * up here by calling {@link ByteString#copyFrom(byte[])} followed by {@link
-   * ByteString#substring(int, int)}.
-   *
-   * <p>This class contains most of the overhead involved in creating a substring
-   * from a {@link LiteralByteString}.  The overhead involves some range-checking
-   * and two extra fields.
-   *
-   * @author carlanton@google.com (Carl Haverl)
-   */
-  // Keep this class private to avoid deadlocks in classloading across threads as ByteString's
-  // static initializer loads LiteralByteString and another thread loads BoundedByteString.
-  private static final class BoundedByteString extends LiteralByteString {
-
-    private final int bytesOffset;
-    private final int bytesLength;
-
-    /**
-     * Creates a {@code BoundedByteString} backed by the sub-range of given array,
-     * without copying.
-     *
-     * @param bytes  array to wrap
-     * @param offset index to first byte to use in bytes
-     * @param length number of bytes to use from bytes
-     * @throws IllegalArgumentException if {@code offset < 0}, {@code length < 0},
-     *                                  or if {@code offset + length >
-     *                                  bytes.length}.
-     */
-    BoundedByteString(byte[] bytes, int offset, int length) {
-      super(bytes);
-      checkRange(offset, offset + length, bytes.length);
-
-      this.bytesOffset = offset;
-      this.bytesLength = length;
-    }
-
-    /**
-     * Gets the byte at the given index.
-     * Throws {@link ArrayIndexOutOfBoundsException}
-     * for backwards-compatibility reasons although it would more properly be
-     * {@link IndexOutOfBoundsException}.
-     *
-     * @param index index of byte
-     * @return the value
-     * @throws ArrayIndexOutOfBoundsException {@code index} is < 0 or >= size
-     */
-    @Override
-    public byte byteAt(int index) {
-      // We must check the index ourselves as we cannot rely on Java array index
-      // checking for substrings.
-      checkIndex(index, size());
-      return bytes[bytesOffset + index];
-    }
-
-    @Override
-    public int size() {
-      return bytesLength;
-    }
-
-    @Override
-    protected int getOffsetIntoBytes() {
-      return bytesOffset;
-    }
-
-    // =================================================================
-    // ByteString -> byte[]
-
-    @Override
-    protected void copyToInternal(byte[] target, int sourceOffset, int targetOffset,
-        int numberToCopy) {
-      System.arraycopy(bytes, getOffsetIntoBytes() + sourceOffset, target,
-          targetOffset, numberToCopy);
-    }
-
-    // =================================================================
-    // Serializable
-
-    private static final long serialVersionUID = 1L;
-
-    Object writeReplace() {
-      return ByteString.wrap(toByteArray());
-    }
-
-    private void readObject(@SuppressWarnings("unused") ObjectInputStream in) throws IOException {
-      throw new InvalidObjectException(
-          "BoundedByteStream instances are not to be serialized directly");
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BytesValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BytesValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BytesValue.java
deleted file mode 100644
index ecfbcc1..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BytesValue.java
+++ /dev/null
@@ -1,454 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `bytes`.
- * The JSON representation for `BytesValue` is JSON string.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.BytesValue}
- */
-public  final class BytesValue extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.BytesValue)
-    BytesValueOrBuilder {
-  // Use BytesValue.newBuilder() to construct.
-  private BytesValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private BytesValue() {
-    value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private BytesValue(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-
-            value_ = input.readBytes();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BytesValue_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BytesValue_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_;
-  /**
-   * <pre>
-   * The bytes value.
-   * </pre>
-   *
-   * <code>bytes value = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!value_.isEmpty()) {
-      output.writeBytes(1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!value_.isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeBytesSize(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue) obj;
-
-    boolean result = true;
-    result = result && getValue()
-        .equals(other.getValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + getValue().hashCode();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `bytes`.
-   * The JSON representation for `BytesValue` is JSON string.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.BytesValue}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.BytesValue)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BytesValue_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BytesValue_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BytesValue_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue.getDefaultInstance()) return this;
-      if (other.getValue() != org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-    /**
-     * <pre>
-     * The bytes value.
-     * </pre>
-     *
-     * <code>bytes value = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * The bytes value.
-     * </pre>
-     *
-     * <code>bytes value = 1;</code>
-     */
-    public Builder setValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The bytes value.
-     * </pre>
-     *
-     * <code>bytes value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = getDefaultInstance().getValue();
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.BytesValue)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.BytesValue)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BytesValue>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<BytesValue>() {
-    public BytesValue parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new BytesValue(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BytesValue> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BytesValue> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.BytesValue getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BytesValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BytesValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BytesValueOrBuilder.java
deleted file mode 100644
index 3f90289..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BytesValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface BytesValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.BytesValue)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The bytes value.
-   * </pre>
-   *
-   * <code>bytes value = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue();
-}


[20/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLiteToString.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLiteToString.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLiteToString.java
deleted file mode 100644
index eea6db5..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageLiteToString.java
+++ /dev/null
@@ -1,239 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.lang.reflect.Method;
-import java.util.HashMap;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-import java.util.Set;
-import java.util.TreeSet;
-
-/**
- * Helps generate {@link String} representations of {@link MessageLite} protos.
- */
-// TODO(dweis): Fix map fields.
-final class MessageLiteToString {
-
-  private static final String LIST_SUFFIX = "List";
-  private static final String BUILDER_LIST_SUFFIX = "OrBuilderList";
-  private static final String BYTES_SUFFIX = "Bytes";
-  
-  /**
-   * Returns a {@link String} representation of the {@link MessageLite} object.  The first line of
-   * the {@code String} representation representation includes a comment string to uniquely identify
-   * the objcet instance. This acts as an indicator that this should not be relied on for
-   * comparisons.
-   *
-   * <p>For use by generated code only.
-   */
-  static String toString(MessageLite messageLite, String commentString) {
-    StringBuilder buffer = new StringBuilder();
-    buffer.append("# ").append(commentString);
-    reflectivePrintWithIndent(messageLite, buffer, 0);
-    return buffer.toString();
-  }
-
-  /**
-   * Reflectively prints the {@link MessageLite} to the buffer at given {@code indent} level.
-   *
-   * @param buffer the buffer to write to
-   * @param indent the number of spaces to indent the proto by
-   */
-  private static void reflectivePrintWithIndent(
-      MessageLite messageLite, StringBuilder buffer, int indent) {
-    // Build a map of method name to method. We're looking for methods like getFoo(), hasFoo(), and
-    // getFooList() which might be useful for building an object's string representation.
-    Map<String, Method> nameToNoArgMethod = new HashMap<String, Method>();
-    Map<String, Method> nameToMethod = new HashMap<String, Method>();
-    Set<String> getters = new TreeSet<String>();
-    for (Method method : messageLite.getClass().getDeclaredMethods()) {
-      nameToMethod.put(method.getName(), method);
-      if (method.getParameterTypes().length == 0) {
-        nameToNoArgMethod.put(method.getName(), method);
-
-        if (method.getName().startsWith("get")) {
-          getters.add(method.getName());
-        }
-      }
-    }
-
-    for (String getter : getters) {
-      String suffix = getter.replaceFirst("get", "");
-      if (suffix.endsWith(LIST_SUFFIX) && !suffix.endsWith(BUILDER_LIST_SUFFIX)) {
-        String camelCase = suffix.substring(0, 1).toLowerCase()
-            + suffix.substring(1, suffix.length() - LIST_SUFFIX.length());
-        // Try to reflectively get the value and toString() the field as if it were repeated. This
-        // only works if the method names have not be proguarded out or renamed.
-        Method listMethod = nameToNoArgMethod.get("get" + suffix);
-        if (listMethod != null && listMethod.getReturnType().equals(List.class)) {
-          printField(
-              buffer,
-              indent,
-              camelCaseToSnakeCase(camelCase),
-              GeneratedMessageLite.invokeOrDie(listMethod, messageLite));
-          continue;
-        }
-      }
-
-      Method setter = nameToMethod.get("set" + suffix);
-      if (setter == null) {
-        continue;
-      }
-      if (suffix.endsWith(BYTES_SUFFIX)
-          && nameToNoArgMethod.containsKey(
-              "get" + suffix.substring(0, suffix.length() - "Bytes".length()))) {
-        // Heuristic to skip bytes based accessors for string fields.
-        continue;
-      }
-
-      String camelCase = suffix.substring(0, 1).toLowerCase() + suffix.substring(1);
-
-      // Try to reflectively get the value and toString() the field as if it were optional. This
-      // only works if the method names have not be proguarded out or renamed.
-      Method getMethod = nameToNoArgMethod.get("get" + suffix);
-      Method hasMethod = nameToNoArgMethod.get("has" + suffix);
-      // TODO(dweis): Fix proto3 semantics.
-      if (getMethod != null) {
-        Object value = GeneratedMessageLite.invokeOrDie(getMethod, messageLite);
-        final boolean hasValue = hasMethod == null
-            ? !isDefaultValue(value)
-            : (Boolean) GeneratedMessageLite.invokeOrDie(hasMethod, messageLite);
-         // TODO(dweis): This doesn't stop printing oneof case twice: value and enum style.
-        if (hasValue) {
-          printField(
-              buffer,
-              indent,
-              camelCaseToSnakeCase(camelCase),
-              value);
-        }
-        continue;
-      }
-    }
-
-    if (messageLite instanceof GeneratedMessageLite.ExtendableMessage) {
-      Iterator<Map.Entry<GeneratedMessageLite.ExtensionDescriptor, Object>> iter =
-          ((GeneratedMessageLite.ExtendableMessage<?, ?>) messageLite).extensions.iterator();
-      while (iter.hasNext()) {
-        Map.Entry<GeneratedMessageLite.ExtensionDescriptor, Object> entry = iter.next();
-        printField(buffer, indent, "[" + entry.getKey().getNumber() + "]", entry.getValue());
-      }
-    }
-
-    if (((GeneratedMessageLite<?, ?>) messageLite).unknownFields != null) {
-      ((GeneratedMessageLite<?, ?>) messageLite).unknownFields.printWithIndent(buffer, indent);
-    }
-  }
-  
-  private static boolean isDefaultValue(Object o) {
-    if (o instanceof Boolean) {
-      return !((Boolean) o);
-    }
-    if (o instanceof Integer) {
-      return ((Integer) o) == 0;
-    }
-    if (o instanceof Float) {
-      return ((Float) o) == 0f;
-    }
-    if (o instanceof Double) {
-      return ((Double) o) == 0d;
-    }
-    if (o instanceof String) {
-      return o.equals("");
-    }
-    if (o instanceof ByteString) {
-      return o.equals(ByteString.EMPTY);
-    }
-    if (o instanceof MessageLite) { // Can happen in oneofs.
-      return o == ((MessageLite) o).getDefaultInstanceForType();
-    }
-    if (o instanceof java.lang.Enum<?>) { // Catches oneof enums.
-      return ((java.lang.Enum<?>) o).ordinal() == 0;
-    }
-    
-    return false;
-  }
-
-  /**
-   * Formats a text proto field.
-   *
-   * <p>For use by generated code only.
-   *
-   * @param buffer the buffer to write to
-   * @param indent the number of spaces the proto should be indented by
-   * @param name the field name (in lower underscore case)
-   * @param object the object value of the field
-   */
-  static final void printField(StringBuilder buffer, int indent, String name, Object object) {
-    if (object instanceof List<?>) {
-      List<?> list = (List<?>) object;
-      for (Object entry : list) {
-        printField(buffer, indent, name, entry);
-      }
-      return;
-    }
-
-    buffer.append('\n');
-    for (int i = 0; i < indent; i++) {
-      buffer.append(' ');
-    }
-    buffer.append(name);
-
-    if (object instanceof String) {
-      buffer.append(": \"").append(TextFormatEscaper.escapeText((String) object)).append('"');
-    } else if (object instanceof ByteString) {
-      buffer.append(": \"").append(TextFormatEscaper.escapeBytes((ByteString) object)).append('"');
-    } else if (object instanceof GeneratedMessageLite) {
-      buffer.append(" {");
-      reflectivePrintWithIndent((GeneratedMessageLite<?, ?>) object, buffer, indent + 2);
-      buffer.append("\n");
-      for (int i = 0; i < indent; i++) {
-        buffer.append(' ');
-      }
-      buffer.append("}");
-    } else {
-      buffer.append(": ").append(object.toString());
-    }
-  }
-  
-  private static final String camelCaseToSnakeCase(String camelCase) {
-    StringBuilder builder = new StringBuilder();
-    for (int i = 0; i < camelCase.length(); i++) {
-      char ch = camelCase.charAt(i);
-      if (Character.isUpperCase(ch)) {
-        builder.append("_");
-      }
-      builder.append(Character.toLowerCase(ch));
-    }
-    return builder.toString();
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageOrBuilder.java
deleted file mode 100644
index 59d5e2f..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageOrBuilder.java
+++ /dev/null
@@ -1,143 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.List;
-import java.util.Map;
-
-/**
- * Base interface for methods common to {@link Message} and
- * {@link Message.Builder} to provide type equivalency.
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public interface MessageOrBuilder extends MessageLiteOrBuilder {
-
-  // (From MessageLite, re-declared here only for return type covariance.)
-  @Override
-  Message getDefaultInstanceForType();
-
-  /**
-   * Returns a list of field paths (e.g. "foo.bar.baz") of required fields
-   * which are not set in this message.  You should call
-   * {@link MessageLiteOrBuilder#isInitialized()} first to check if there
-   * are any missing fields, as that method is likely to be much faster
-   * than this one even when the message is fully-initialized.
-   */
-  List<String> findInitializationErrors();
-
-  /**
-   * Returns a comma-delimited list of required fields which are not set
-   * in this message object.  You should call
-   * {@link MessageLiteOrBuilder#isInitialized()} first to check if there
-   * are any missing fields, as that method is likely to be much faster
-   * than this one even when the message is fully-initialized.
-   */
-  String getInitializationErrorString();
-
-  /**
-   * Get the message's type's descriptor.  This differs from the
-   * {@code getDescriptor()} method of generated message classes in that
-   * this method is an abstract method of the {@code Message} interface
-   * whereas {@code getDescriptor()} is a static method of a specific class.
-   * They return the same thing.
-   */
-  Descriptors.Descriptor getDescriptorForType();
-
-  /**
-   * Returns a collection of all the fields in this message which are set
-   * and their corresponding values.  A singular ("required" or "optional")
-   * field is set iff hasField() returns true for that field.  A "repeated"
-   * field is set iff getRepeatedFieldCount() is greater than zero.  The
-   * values are exactly what would be returned by calling
-   * {@link #getField(Descriptors.FieldDescriptor)} for each field.  The map
-   * is guaranteed to be a sorted map, so iterating over it will return fields
-   * in order by field number.
-   * <br>
-   * If this is for a builder, the returned map may or may not reflect future
-   * changes to the builder.  Either way, the returned map is itself
-   * unmodifiable.
-   */
-  Map<Descriptors.FieldDescriptor, Object> getAllFields();
-
-  /**
-   * Returns true if the given oneof is set.
-   * @throws IllegalArgumentException if
-   *           {@code oneof.getContainingType() != getDescriptorForType()}.
-   */
-  boolean hasOneof(Descriptors.OneofDescriptor oneof);
-
-  /**
-   * Obtains the FieldDescriptor if the given oneof is set. Returns null
-   * if no field is set.
-   */
-  Descriptors.FieldDescriptor getOneofFieldDescriptor(
-      Descriptors.OneofDescriptor oneof);
-
-  /**
-   * Returns true if the given field is set.  This is exactly equivalent to
-   * calling the generated "has" accessor method corresponding to the field.
-   * @throws IllegalArgumentException The field is a repeated field, or
-   *           {@code field.getContainingType() != getDescriptorForType()}.
-   */
-  boolean hasField(Descriptors.FieldDescriptor field);
-
-  /**
-   * Obtains the value of the given field, or the default value if it is
-   * not set.  For primitive fields, the boxed primitive value is returned.
-   * For enum fields, the EnumValueDescriptor for the value is returned. For
-   * embedded message fields, the sub-message is returned.  For repeated
-   * fields, a java.util.List is returned.
-   */
-  Object getField(Descriptors.FieldDescriptor field);
-
-  /**
-   * Gets the number of elements of a repeated field.  This is exactly
-   * equivalent to calling the generated "Count" accessor method corresponding
-   * to the field.
-   * @throws IllegalArgumentException The field is not a repeated field, or
-   *           {@code field.getContainingType() != getDescriptorForType()}.
-   */
-  int getRepeatedFieldCount(Descriptors.FieldDescriptor field);
-
-  /**
-   * Gets an element of a repeated field.  For primitive fields, the boxed
-   * primitive value is returned.  For enum fields, the EnumValueDescriptor
-   * for the value is returned. For embedded message fields, the sub-message
-   * is returned.
-   * @throws IllegalArgumentException The field is not a repeated field, or
-   *           {@code field.getContainingType() != getDescriptorForType()}.
-   */
-  Object getRepeatedField(Descriptors.FieldDescriptor field, int index);
-
-  /** Get the {@link UnknownFieldSet} for this message. */
-  UnknownFieldSet getUnknownFields();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageReflection.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageReflection.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageReflection.java
deleted file mode 100644
index 99a3377..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MessageReflection.java
+++ /dev/null
@@ -1,990 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-
-import java.io.IOException;
-import java.util.ArrayList;
-import java.util.List;
-import java.util.Map;
-import java.util.TreeMap;
-
-/**
- * Reflection utility methods shared by both mutable and immutable messages.
- *
- * @author liujisi@google.com (Pherl Liu)
- */
-class MessageReflection {
-
-  static void writeMessageTo(
-      Message message,
-      Map<FieldDescriptor, Object> fields,
-      CodedOutputStream output,
-      boolean alwaysWriteRequiredFields)
-      throws IOException {
-    final boolean isMessageSet =
-        message.getDescriptorForType().getOptions().getMessageSetWireFormat();
-    if (alwaysWriteRequiredFields) {
-      fields = new TreeMap<FieldDescriptor, Object>(fields);
-      for (final FieldDescriptor field :
-          message.getDescriptorForType().getFields()) {
-        if (field.isRequired() && !fields.containsKey(field)) {
-          fields.put(field, message.getField(field));
-        }
-      }
-    }
-    for (final Map.Entry<Descriptors.FieldDescriptor, Object> entry :
-        fields.entrySet()) {
-      final Descriptors.FieldDescriptor field = entry.getKey();
-      final Object value = entry.getValue();
-      if (isMessageSet && field.isExtension() &&
-          field.getType() == Descriptors.FieldDescriptor.Type.MESSAGE &&
-          !field.isRepeated()) {
-        output.writeMessageSetExtension(field.getNumber(), (Message) value);
-      } else {
-        FieldSet.writeField(field, value, output);
-      }
-    }
-
-    final UnknownFieldSet unknownFields = message.getUnknownFields();
-    if (isMessageSet) {
-      unknownFields.writeAsMessageSetTo(output);
-    } else {
-      unknownFields.writeTo(output);
-    }
-  }
-
-  static int getSerializedSize(
-      Message message,
-      Map<FieldDescriptor, Object> fields) {
-    int size = 0;
-    final boolean isMessageSet =
-        message.getDescriptorForType().getOptions().getMessageSetWireFormat();
-
-    for (final Map.Entry<Descriptors.FieldDescriptor, Object> entry :
-        fields.entrySet()) {
-      final Descriptors.FieldDescriptor field = entry.getKey();
-      final Object value = entry.getValue();
-      if (isMessageSet && field.isExtension() &&
-          field.getType() == Descriptors.FieldDescriptor.Type.MESSAGE &&
-          !field.isRepeated()) {
-        size += CodedOutputStream.computeMessageSetExtensionSize(
-            field.getNumber(), (Message) value);
-      } else {
-        size += FieldSet.computeFieldSize(field, value);
-      }
-    }
-
-    final UnknownFieldSet unknownFields = message.getUnknownFields();
-    if (isMessageSet) {
-      size += unknownFields.getSerializedSizeAsMessageSet();
-    } else {
-      size += unknownFields.getSerializedSize();
-    }
-    return size;
-  }
-
-  static String delimitWithCommas(List<String> parts) {
-    StringBuilder result = new StringBuilder();
-    for (String part : parts) {
-      if (result.length() > 0) {
-        result.append(", ");
-      }
-      result.append(part);
-    }
-    return result.toString();
-  }
-
-  @SuppressWarnings("unchecked")
-  static boolean isInitialized(MessageOrBuilder message) {
-    // Check that all required fields are present.
-    for (final Descriptors.FieldDescriptor field : message
-        .getDescriptorForType()
-        .getFields()) {
-      if (field.isRequired()) {
-        if (!message.hasField(field)) {
-          return false;
-        }
-      }
-    }
-
-    // Check that embedded messages are initialized.
-    for (final Map.Entry<Descriptors.FieldDescriptor, Object> entry :
-        message.getAllFields().entrySet()) {
-      final Descriptors.FieldDescriptor field = entry.getKey();
-      if (field.getJavaType() == Descriptors.FieldDescriptor.JavaType.MESSAGE) {
-        if (field.isRepeated()) {
-          for (final Message element
-              : (List<Message>) entry.getValue()) {
-            if (!element.isInitialized()) {
-              return false;
-            }
-          }
-        } else {
-          if (!((Message) entry.getValue()).isInitialized()) {
-            return false;
-          }
-        }
-      }
-    }
-
-    return true;
-  }
-
-  private static String subMessagePrefix(final String prefix,
-      final Descriptors.FieldDescriptor field,
-      final int index) {
-    final StringBuilder result = new StringBuilder(prefix);
-    if (field.isExtension()) {
-      result.append('(')
-          .append(field.getFullName())
-          .append(')');
-    } else {
-      result.append(field.getName());
-    }
-    if (index != -1) {
-      result.append('[')
-          .append(index)
-          .append(']');
-    }
-    result.append('.');
-    return result.toString();
-  }
-
-  private static void findMissingFields(final MessageOrBuilder message,
-      final String prefix,
-      final List<String> results) {
-    for (final Descriptors.FieldDescriptor field :
-        message.getDescriptorForType().getFields()) {
-      if (field.isRequired() && !message.hasField(field)) {
-        results.add(prefix + field.getName());
-      }
-    }
-
-    for (final Map.Entry<Descriptors.FieldDescriptor, Object> entry :
-        message.getAllFields().entrySet()) {
-      final Descriptors.FieldDescriptor field = entry.getKey();
-      final Object value = entry.getValue();
-
-      if (field.getJavaType() == Descriptors.FieldDescriptor.JavaType.MESSAGE) {
-        if (field.isRepeated()) {
-          int i = 0;
-          for (final Object element : (List) value) {
-            findMissingFields((MessageOrBuilder) element,
-                subMessagePrefix(prefix, field, i++),
-                results);
-          }
-        } else {
-          if (message.hasField(field)) {
-            findMissingFields((MessageOrBuilder) value,
-                subMessagePrefix(prefix, field, -1),
-                results);
-          }
-        }
-      }
-    }
-  }
-
-  /**
-   * Populates {@code this.missingFields} with the full "path" of each missing
-   * required field in the given message.
-   */
-  static List<String> findMissingFields(
-      final MessageOrBuilder message) {
-    final List<String> results = new ArrayList<String>();
-    findMissingFields(message, "", results);
-    return results;
-  }
-
-  static interface MergeTarget {
-    enum ContainerType {
-      MESSAGE, EXTENSION_SET
-    }
-
-    /**
-     * Returns the descriptor for the target.
-     */
-    public Descriptors.Descriptor getDescriptorForType();
-
-    public ContainerType getContainerType();
-
-    public ExtensionRegistry.ExtensionInfo findExtensionByName(
-        ExtensionRegistry registry, String name);
-
-    public ExtensionRegistry.ExtensionInfo findExtensionByNumber(
-        ExtensionRegistry registry, Descriptors.Descriptor containingType,
-        int fieldNumber);
-
-    /**
-     * Obtains the value of the given field, or the default value if it is not
-     * set.  For primitive fields, the boxed primitive value is returned. For
-     * enum fields, the EnumValueDescriptor for the value is returned. For
-     * embedded message fields, the sub-message is returned.  For repeated
-     * fields, a java.util.List is returned.
-     */
-    public Object getField(Descriptors.FieldDescriptor field);
-
-    /**
-     * Returns true if the given field is set.  This is exactly equivalent to
-     * calling the generated "has" accessor method corresponding to the field.
-     *
-     * @throws IllegalArgumentException The field is a repeated field, or {@code
-     *     field.getContainingType() != getDescriptorForType()}.
-     */
-    boolean hasField(Descriptors.FieldDescriptor field);
-
-    /**
-     * Sets a field to the given value.  The value must be of the correct type
-     * for this field, i.e. the same type that
-     * {@link Message#getField(Descriptors.FieldDescriptor)}
-     * would return.
-     */
-    MergeTarget setField(Descriptors.FieldDescriptor field, Object value);
-
-    /**
-     * Clears the field.  This is exactly equivalent to calling the generated
-     * "clear" accessor method corresponding to the field.
-     */
-    MergeTarget clearField(Descriptors.FieldDescriptor field);
-
-    /**
-     * Sets an element of a repeated field to the given value.  The value must
-     * be of the correct type for this field, i.e. the same type that {@link
-     * Message#getRepeatedField(Descriptors.FieldDescriptor, int)} would return.
-     *
-     * @throws IllegalArgumentException The field is not a repeated field, or
-     *                                  {@code field.getContainingType() !=
-     *                                  getDescriptorForType()}.
-     */
-    MergeTarget setRepeatedField(Descriptors.FieldDescriptor field,
-        int index, Object value);
-
-    /**
-     * Like {@code setRepeatedField}, but appends the value as a new element.
-     *
-     * @throws IllegalArgumentException The field is not a repeated field, or
-     *                                  {@code field.getContainingType() !=
-     *                                  getDescriptorForType()}.
-     */
-    MergeTarget addRepeatedField(Descriptors.FieldDescriptor field,
-        Object value);
-
-    /**
-     * Returns true if the given oneof is set.
-     *
-     * @throws IllegalArgumentException if
-     *           {@code oneof.getContainingType() != getDescriptorForType()}.
-     */
-    boolean hasOneof(Descriptors.OneofDescriptor oneof);
-
-    /**
-     * Clears the oneof.  This is exactly equivalent to calling the generated
-     * "clear" accessor method corresponding to the oneof.
-     */
-    MergeTarget clearOneof(Descriptors.OneofDescriptor oneof);
-
-    /**
-     * Obtains the FieldDescriptor if the given oneof is set. Returns null
-     * if no field is set.
-     */
-    Descriptors.FieldDescriptor getOneofFieldDescriptor(Descriptors.OneofDescriptor oneof);
-
-    /**
-     * Parse the input stream into a sub field group defined based on either
-     * FieldDescriptor or the default instance.
-     */
-    Object parseGroup(CodedInputStream input, ExtensionRegistryLite registry,
-        Descriptors.FieldDescriptor descriptor, Message defaultInstance)
-        throws IOException;
-
-    /**
-     * Parse the input stream into a sub field message defined based on either
-     * FieldDescriptor or the default instance.
-     */
-    Object parseMessage(CodedInputStream input, ExtensionRegistryLite registry,
-        Descriptors.FieldDescriptor descriptor, Message defaultInstance)
-        throws IOException;
-
-    /**
-     * Parse from a ByteString into a sub field message defined based on either
-     * FieldDescriptor or the default instance.  There isn't a varint indicating
-     * the length of the message at the beginning of the input ByteString.
-     */
-    Object parseMessageFromBytes(
-        ByteString bytes, ExtensionRegistryLite registry,
-        Descriptors.FieldDescriptor descriptor, Message defaultInstance)
-        throws IOException;
-
-    /**
-     * Returns the UTF8 validation level for the field.
-     */
-    WireFormat.Utf8Validation getUtf8Validation(Descriptors.FieldDescriptor
-        descriptor);
-
-    /**
-     * Returns a new merge target for a sub-field. When defaultInstance is
-     * provided, it indicates the descriptor is for an extension type, and
-     * implementations should create a new instance from the defaultInstance
-     * prototype directly.
-     */
-    MergeTarget newMergeTargetForField(
-        Descriptors.FieldDescriptor descriptor,
-        Message defaultInstance);
-
-    /**
-     * Finishes the merge and returns the underlying object.
-     */
-    Object finish();
-  }
-
-  static class BuilderAdapter implements MergeTarget {
-
-    private final Message.Builder builder;
-
-    @Override
-    public Descriptors.Descriptor getDescriptorForType() {
-      return builder.getDescriptorForType();
-    }
-
-    public BuilderAdapter(Message.Builder builder) {
-      this.builder = builder;
-    }
-
-    @Override
-    public Object getField(Descriptors.FieldDescriptor field) {
-      return builder.getField(field);
-    }
-
-    @Override
-    public boolean hasField(Descriptors.FieldDescriptor field) {
-      return builder.hasField(field);
-    }
-
-    @Override
-    public MergeTarget setField(Descriptors.FieldDescriptor field, Object value) {
-      builder.setField(field, value);
-      return this;
-    }
-
-    @Override
-    public MergeTarget clearField(Descriptors.FieldDescriptor field) {
-      builder.clearField(field);
-      return this;
-    }
-
-    @Override
-    public MergeTarget setRepeatedField(
-        Descriptors.FieldDescriptor field, int index, Object value) {
-      builder.setRepeatedField(field, index, value);
-      return this;
-    }
-
-    @Override
-    public MergeTarget addRepeatedField(Descriptors.FieldDescriptor field, Object value) {
-      builder.addRepeatedField(field, value);
-      return this;
-    }
-
-    @Override
-    public boolean hasOneof(Descriptors.OneofDescriptor oneof) {
-      return builder.hasOneof(oneof);
-    }
-
-    @Override
-    public MergeTarget clearOneof(Descriptors.OneofDescriptor oneof) {
-      builder.clearOneof(oneof);
-      return this;
-    }
-
-    @Override
-    public Descriptors.FieldDescriptor getOneofFieldDescriptor(Descriptors.OneofDescriptor oneof) {
-      return builder.getOneofFieldDescriptor(oneof);
-    }
-
-    @Override
-    public ContainerType getContainerType() {
-      return ContainerType.MESSAGE;
-    }
-
-    @Override
-    public ExtensionRegistry.ExtensionInfo findExtensionByName(
-        ExtensionRegistry registry, String name) {
-      return registry.findImmutableExtensionByName(name);
-    }
-
-    @Override
-    public ExtensionRegistry.ExtensionInfo findExtensionByNumber(
-        ExtensionRegistry registry, Descriptors.Descriptor containingType, int fieldNumber) {
-      return registry.findImmutableExtensionByNumber(containingType,
-          fieldNumber);
-    }
-
-    @Override
-    public Object parseGroup(
-        CodedInputStream input,
-        ExtensionRegistryLite extensionRegistry,
-        Descriptors.FieldDescriptor field,
-        Message defaultInstance)
-        throws IOException {
-      Message.Builder subBuilder;
-      // When default instance is not null. The field is an extension field.
-      if (defaultInstance != null) {
-        subBuilder = defaultInstance.newBuilderForType();
-      } else {
-        subBuilder = builder.newBuilderForField(field);
-      }
-      if (!field.isRepeated()) {
-        Message originalMessage = (Message) getField(field);
-        if (originalMessage != null) {
-          subBuilder.mergeFrom(originalMessage);
-        }
-      }
-      input.readGroup(field.getNumber(), subBuilder, extensionRegistry);
-      return subBuilder.buildPartial();
-    }
-
-    @Override
-    public Object parseMessage(
-        CodedInputStream input,
-        ExtensionRegistryLite extensionRegistry,
-        Descriptors.FieldDescriptor field,
-        Message defaultInstance)
-        throws IOException {
-      Message.Builder subBuilder;
-      // When default instance is not null. The field is an extension field.
-      if (defaultInstance != null) {
-        subBuilder = defaultInstance.newBuilderForType();
-      } else {
-        subBuilder = builder.newBuilderForField(field);
-      }
-      if (!field.isRepeated()) {
-        Message originalMessage = (Message) getField(field);
-        if (originalMessage != null) {
-          subBuilder.mergeFrom(originalMessage);
-        }
-      }
-      input.readMessage(subBuilder, extensionRegistry);
-      return subBuilder.buildPartial();
-    }
-
-    @Override
-    public Object parseMessageFromBytes(
-        ByteString bytes,
-        ExtensionRegistryLite extensionRegistry,
-        Descriptors.FieldDescriptor field,
-        Message defaultInstance)
-        throws IOException {
-      Message.Builder subBuilder;
-      // When default instance is not null. The field is an extension field.
-      if (defaultInstance != null) {
-        subBuilder = defaultInstance.newBuilderForType();
-      } else {
-        subBuilder = builder.newBuilderForField(field);
-      }
-      if (!field.isRepeated()) {
-        Message originalMessage = (Message) getField(field);
-        if (originalMessage != null) {
-          subBuilder.mergeFrom(originalMessage);
-        }
-      }
-      subBuilder.mergeFrom(bytes, extensionRegistry);
-      return subBuilder.buildPartial();
-    }
-
-    @Override
-    public MergeTarget newMergeTargetForField(
-        Descriptors.FieldDescriptor field, Message defaultInstance) {
-      if (defaultInstance != null) {
-        return new BuilderAdapter(
-            defaultInstance.newBuilderForType());
-      } else {
-        return new BuilderAdapter(builder.newBuilderForField(field));
-      }
-    }
-
-    @Override
-    public WireFormat.Utf8Validation getUtf8Validation(Descriptors.FieldDescriptor descriptor) {
-      if (descriptor.needsUtf8Check()) {
-        return WireFormat.Utf8Validation.STRICT;
-      }
-      // TODO(liujisi): support lazy strings for repeated fields.
-      if (!descriptor.isRepeated()
-          && builder instanceof GeneratedMessage.Builder) {
-        return WireFormat.Utf8Validation.LAZY;
-      }
-      return WireFormat.Utf8Validation.LOOSE;
-    }
-
-    @Override
-    public Object finish() {
-      return builder.buildPartial();
-    }
-  }
-
-
-  static class ExtensionAdapter implements MergeTarget {
-
-    private final FieldSet<Descriptors.FieldDescriptor> extensions;
-
-    ExtensionAdapter(FieldSet<Descriptors.FieldDescriptor> extensions) {
-      this.extensions = extensions;
-    }
-
-    @Override
-    public Descriptors.Descriptor getDescriptorForType() {
-      throw new UnsupportedOperationException(
-          "getDescriptorForType() called on FieldSet object");
-    }
-
-    @Override
-    public Object getField(Descriptors.FieldDescriptor field) {
-      return extensions.getField(field);
-    }
-
-    @Override
-    public boolean hasField(Descriptors.FieldDescriptor field) {
-      return extensions.hasField(field);
-    }
-
-    @Override
-    public MergeTarget setField(Descriptors.FieldDescriptor field, Object value) {
-      extensions.setField(field, value);
-      return this;
-    }
-
-    @Override
-    public MergeTarget clearField(Descriptors.FieldDescriptor field) {
-      extensions.clearField(field);
-      return this;
-    }
-
-    @Override
-    public MergeTarget setRepeatedField(
-        Descriptors.FieldDescriptor field, int index, Object value) {
-      extensions.setRepeatedField(field, index, value);
-      return this;
-    }
-
-    @Override
-    public MergeTarget addRepeatedField(Descriptors.FieldDescriptor field, Object value) {
-      extensions.addRepeatedField(field, value);
-      return this;
-    }
-
-    @Override
-    public boolean hasOneof(Descriptors.OneofDescriptor oneof) {
-      return false;
-    }
-
-    @Override
-    public MergeTarget clearOneof(Descriptors.OneofDescriptor oneof) {
-      // Nothing to clear.
-      return this;
-    }
-
-    @Override
-    public Descriptors.FieldDescriptor getOneofFieldDescriptor(Descriptors.OneofDescriptor oneof) {
-      return null;
-    }
-
-    @Override
-    public ContainerType getContainerType() {
-      return ContainerType.EXTENSION_SET;
-    }
-
-    @Override
-    public ExtensionRegistry.ExtensionInfo findExtensionByName(
-        ExtensionRegistry registry, String name) {
-      return registry.findImmutableExtensionByName(name);
-    }
-
-    @Override
-    public ExtensionRegistry.ExtensionInfo findExtensionByNumber(
-        ExtensionRegistry registry, Descriptors.Descriptor containingType, int fieldNumber) {
-      return registry.findImmutableExtensionByNumber(containingType,
-          fieldNumber);
-    }
-
-    @Override
-    public Object parseGroup(
-        CodedInputStream input,
-        ExtensionRegistryLite registry,
-        Descriptors.FieldDescriptor field,
-        Message defaultInstance)
-        throws IOException {
-      Message.Builder subBuilder =
-          defaultInstance.newBuilderForType();
-      if (!field.isRepeated()) {
-        Message originalMessage = (Message) getField(field);
-        if (originalMessage != null) {
-          subBuilder.mergeFrom(originalMessage);
-        }
-      }
-      input.readGroup(field.getNumber(), subBuilder, registry);
-      return subBuilder.buildPartial();
-    }
-
-    @Override
-    public Object parseMessage(
-        CodedInputStream input,
-        ExtensionRegistryLite registry,
-        Descriptors.FieldDescriptor field,
-        Message defaultInstance)
-        throws IOException {
-      Message.Builder subBuilder =
-          defaultInstance.newBuilderForType();
-      if (!field.isRepeated()) {
-        Message originalMessage = (Message) getField(field);
-        if (originalMessage != null) {
-          subBuilder.mergeFrom(originalMessage);
-        }
-      }
-      input.readMessage(subBuilder, registry);
-      return subBuilder.buildPartial();
-    }
-
-    @Override
-    public Object parseMessageFromBytes(
-        ByteString bytes,
-        ExtensionRegistryLite registry,
-        Descriptors.FieldDescriptor field,
-        Message defaultInstance)
-        throws IOException {
-      Message.Builder subBuilder =  defaultInstance.newBuilderForType();
-      if (!field.isRepeated()) {
-        Message originalMessage = (Message) getField(field);
-        if (originalMessage != null) {
-          subBuilder.mergeFrom(originalMessage);
-        }
-      }
-      subBuilder.mergeFrom(bytes, registry);
-      return subBuilder.buildPartial();
-    }
-
-    @Override
-    public MergeTarget newMergeTargetForField(
-        Descriptors.FieldDescriptor descriptor, Message defaultInstance) {
-      throw new UnsupportedOperationException(
-          "newMergeTargetForField() called on FieldSet object");
-    }
-
-    @Override
-    public WireFormat.Utf8Validation getUtf8Validation(Descriptors.FieldDescriptor descriptor) {
-      if (descriptor.needsUtf8Check()) {
-        return WireFormat.Utf8Validation.STRICT;
-      }
-      // TODO(liujisi): support lazy strings for ExtesnsionSet.
-      return WireFormat.Utf8Validation.LOOSE;
-    }
-
-    @Override
-    public Object finish() {
-      throw new UnsupportedOperationException(
-          "finish() called on FieldSet object");
-    }
-  }
-
-  /**
-   * Parses a single field into MergeTarget. The target can be Message.Builder,
-   * FieldSet or MutableMessage.
-   *
-   * Package-private because it is used by GeneratedMessage.ExtendableMessage.
-   *
-   * @param tag The tag, which should have already been read.
-   * @return {@code true} unless the tag is an end-group tag.
-   */
-  static boolean mergeFieldFrom(
-      CodedInputStream input,
-      UnknownFieldSet.Builder unknownFields,
-      ExtensionRegistryLite extensionRegistry,
-      Descriptors.Descriptor type,
-      MergeTarget target,
-      int tag) throws IOException {
-    if (type.getOptions().getMessageSetWireFormat() &&
-        tag == WireFormat.MESSAGE_SET_ITEM_TAG) {
-      mergeMessageSetExtensionFromCodedStream(
-          input, unknownFields, extensionRegistry, type, target);
-      return true;
-    }
-
-    final int wireType = WireFormat.getTagWireType(tag);
-    final int fieldNumber = WireFormat.getTagFieldNumber(tag);
-
-    final Descriptors.FieldDescriptor field;
-    Message defaultInstance = null;
-
-    if (type.isExtensionNumber(fieldNumber)) {
-      // extensionRegistry may be either ExtensionRegistry or
-      // ExtensionRegistryLite.  Since the type we are parsing is a full
-      // message, only a full ExtensionRegistry could possibly contain
-      // extensions of it.  Otherwise we will treat the registry as if it
-      // were empty.
-      if (extensionRegistry instanceof ExtensionRegistry) {
-        final ExtensionRegistry.ExtensionInfo extension =
-            target.findExtensionByNumber((ExtensionRegistry) extensionRegistry,
-                type, fieldNumber);
-        if (extension == null) {
-          field = null;
-        } else {
-          field = extension.descriptor;
-          defaultInstance = extension.defaultInstance;
-          if (defaultInstance == null &&
-              field.getJavaType()
-                  == Descriptors.FieldDescriptor.JavaType.MESSAGE) {
-            throw new IllegalStateException(
-                "Message-typed extension lacked default instance: " +
-                    field.getFullName());
-          }
-        }
-      } else {
-        field = null;
-      }
-    } else if (target.getContainerType() == MergeTarget.ContainerType.MESSAGE) {
-      field = type.findFieldByNumber(fieldNumber);
-    } else {
-      field = null;
-    }
-
-    boolean unknown = false;
-    boolean packed = false;
-    if (field == null) {
-      unknown = true;  // Unknown field.
-    } else if (wireType == FieldSet.getWireFormatForFieldType(
-        field.getLiteType(),
-        false  /* isPacked */)) {
-      packed = false;
-    } else if (field.isPackable() &&
-        wireType == FieldSet.getWireFormatForFieldType(
-            field.getLiteType(),
-            true  /* isPacked */)) {
-      packed = true;
-    } else {
-      unknown = true;  // Unknown wire type.
-    }
-
-    if (unknown) {  // Unknown field or wrong wire type.  Skip.
-      return unknownFields.mergeFieldFrom(tag, input);
-    }
-
-    if (packed) {
-      final int length = input.readRawVarint32();
-      final int limit = input.pushLimit(length);
-      if (field.getLiteType() == WireFormat.FieldType.ENUM) {
-        while (input.getBytesUntilLimit() > 0) {
-          final int rawValue = input.readEnum();
-          if (field.getFile().supportsUnknownEnumValue()) {
-            target.addRepeatedField(field,
-                field.getEnumType().findValueByNumberCreatingIfUnknown(rawValue));
-          } else {
-            final Object value = field.getEnumType().findValueByNumber(rawValue);
-            if (value == null) {
-              // If the number isn't recognized as a valid value for this
-              // enum, drop it (don't even add it to unknownFields).
-              return true;
-            }
-            target.addRepeatedField(field, value);
-          }
-        }
-      } else {
-        while (input.getBytesUntilLimit() > 0) {
-          final Object value = WireFormat.readPrimitiveField(
-              input, field.getLiteType(), target.getUtf8Validation(field));
-          target.addRepeatedField(field, value);
-        }
-      }
-      input.popLimit(limit);
-    } else {
-      final Object value;
-      switch (field.getType()) {
-        case GROUP: {
-          value = target
-              .parseGroup(input, extensionRegistry, field, defaultInstance);
-          break;
-        }
-        case MESSAGE: {
-          value = target
-              .parseMessage(input, extensionRegistry, field, defaultInstance);
-          break;
-        }
-        case ENUM:
-          final int rawValue = input.readEnum();
-          if (field.getFile().supportsUnknownEnumValue()) {
-            value = field.getEnumType().findValueByNumberCreatingIfUnknown(rawValue);
-          } else {
-            value = field.getEnumType().findValueByNumber(rawValue);
-            // If the number isn't recognized as a valid value for this enum,
-            // drop it.
-            if (value == null) {
-              unknownFields.mergeVarintField(fieldNumber, rawValue);
-              return true;
-            }
-          }
-          break;
-        default:
-          value = WireFormat.readPrimitiveField(
-              input, field.getLiteType(), target.getUtf8Validation(field));
-          break;
-      }
-
-      if (field.isRepeated()) {
-        target.addRepeatedField(field, value);
-      } else {
-        target.setField(field, value);
-      }
-    }
-
-    return true;
-  }
-
-  /**
-   * Called by {@code #mergeFieldFrom()} to parse a MessageSet extension into
-   * MergeTarget.
-   */
-  private static void mergeMessageSetExtensionFromCodedStream(
-      CodedInputStream input,
-      UnknownFieldSet.Builder unknownFields,
-      ExtensionRegistryLite extensionRegistry,
-      Descriptors.Descriptor type,
-      MergeTarget target) throws IOException {
-
-    // The wire format for MessageSet is:
-    //   message MessageSet {
-    //     repeated group Item = 1 {
-    //       required int32 typeId = 2;
-    //       required bytes message = 3;
-    //     }
-    //   }
-    // "typeId" is the extension's field number.  The extension can only be
-    // a message type, where "message" contains the encoded bytes of that
-    // message.
-    //
-    // In practice, we will probably never see a MessageSet item in which
-    // the message appears before the type ID, or where either field does not
-    // appear exactly once.  However, in theory such cases are valid, so we
-    // should be prepared to accept them.
-
-    int typeId = 0;
-    ByteString rawBytes = null; // If we encounter "message" before "typeId"
-    ExtensionRegistry.ExtensionInfo extension = null;
-
-    // Read bytes from input, if we get it's type first then parse it eagerly,
-    // otherwise we store the raw bytes in a local variable.
-    while (true) {
-      final int tag = input.readTag();
-      if (tag == 0) {
-        break;
-      }
-
-      if (tag == WireFormat.MESSAGE_SET_TYPE_ID_TAG) {
-        typeId = input.readUInt32();
-        if (typeId != 0) {
-          // extensionRegistry may be either ExtensionRegistry or
-          // ExtensionRegistryLite. Since the type we are parsing is a full
-          // message, only a full ExtensionRegistry could possibly contain
-          // extensions of it. Otherwise we will treat the registry as if it
-          // were empty.
-          if (extensionRegistry instanceof ExtensionRegistry) {
-            extension = target.findExtensionByNumber(
-                (ExtensionRegistry) extensionRegistry, type, typeId);
-          }
-        }
-
-      } else if (tag == WireFormat.MESSAGE_SET_MESSAGE_TAG) {
-        if (typeId != 0) {
-          if (extension != null &&
-              ExtensionRegistryLite.isEagerlyParseMessageSets()) {
-            // We already know the type, so we can parse directly from the
-            // input with no copying.  Hooray!
-            eagerlyMergeMessageSetExtension(
-                input, extension, extensionRegistry, target);
-            rawBytes = null;
-            continue;
-          }
-        }
-        // We haven't seen a type ID yet or we want parse message lazily.
-        rawBytes = input.readBytes();
-
-      } else { // Unknown tag. Skip it.
-        if (!input.skipField(tag)) {
-          break; // End of group
-        }
-      }
-    }
-    input.checkLastTagWas(WireFormat.MESSAGE_SET_ITEM_END_TAG);
-
-    // Process the raw bytes.
-    if (rawBytes != null && typeId != 0) { // Zero is not a valid type ID.
-      if (extension != null) { // We known the type
-        mergeMessageSetExtensionFromBytes(
-            rawBytes, extension, extensionRegistry, target);
-      } else { // We don't know how to parse this. Ignore it.
-        if (rawBytes != null) {
-          unknownFields.mergeField(typeId, UnknownFieldSet.Field.newBuilder()
-              .addLengthDelimited(rawBytes).build());
-        }
-      }
-    }
-  }
-
-  private static void mergeMessageSetExtensionFromBytes(
-      ByteString rawBytes,
-      ExtensionRegistry.ExtensionInfo extension,
-      ExtensionRegistryLite extensionRegistry,
-      MergeTarget target) throws IOException {
-
-    Descriptors.FieldDescriptor field = extension.descriptor;
-    boolean hasOriginalValue = target.hasField(field);
-
-    if (hasOriginalValue || ExtensionRegistryLite.isEagerlyParseMessageSets()) {
-      // If the field already exists, we just parse the field.
-      Object value = target.parseMessageFromBytes(
-          rawBytes, extensionRegistry,field, extension.defaultInstance);
-      target.setField(field, value);
-    } else {
-      // Use LazyField to load MessageSet lazily.
-      LazyField lazyField = new LazyField(
-          extension.defaultInstance, extensionRegistry, rawBytes);
-      target.setField(field, lazyField);
-    }
-  }
-
-  private static void eagerlyMergeMessageSetExtension(
-      CodedInputStream input,
-      ExtensionRegistry.ExtensionInfo extension,
-      ExtensionRegistryLite extensionRegistry,
-      MergeTarget target) throws IOException {
-    Descriptors.FieldDescriptor field = extension.descriptor;
-    Object value = target.parseMessage(input, extensionRegistry, field,
-                                       extension.defaultInstance);
-    target.setField(field, value);
-  }
-}


[47/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithNullResponseProtos.java
----------------------------------------------------------------------
diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithNullResponseProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithNullResponseProtos.java
deleted file mode 100644
index b25f7aa..0000000
--- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithNullResponseProtos.java
+++ /dev/null
@@ -1,1283 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: ColumnAggregationNullResponseProtocol.proto
-
-package org.apache.hadoop.hbase.coprocessor.protobuf.generated;
-
-public final class ColumnAggregationWithNullResponseProtos {
-  private ColumnAggregationWithNullResponseProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface ColumnAggregationNullResponseSumRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required bytes family = 1;
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    boolean hasFamily();
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    com.google.protobuf.ByteString getFamily();
-
-    // optional bytes qualifier = 2;
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    boolean hasQualifier();
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    com.google.protobuf.ByteString getQualifier();
-  }
-  /**
-   * Protobuf type {@code ColumnAggregationNullResponseSumRequest}
-   *
-   * <pre>
-   * use unique names for messages in ColumnAggregationXXX.protos due to a bug in
-   * protoc or hadoop's protoc compiler.
-   * </pre>
-   */
-  public static final class ColumnAggregationNullResponseSumRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements ColumnAggregationNullResponseSumRequestOrBuilder {
-    // Use ColumnAggregationNullResponseSumRequest.newBuilder() to construct.
-    private ColumnAggregationNullResponseSumRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private ColumnAggregationNullResponseSumRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final ColumnAggregationNullResponseSumRequest defaultInstance;
-    public static ColumnAggregationNullResponseSumRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public ColumnAggregationNullResponseSumRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private ColumnAggregationNullResponseSumRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              family_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              bitField0_ |= 0x00000002;
-              qualifier_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<ColumnAggregationNullResponseSumRequest> PARSER =
-        new com.google.protobuf.AbstractParser<ColumnAggregationNullResponseSumRequest>() {
-      public ColumnAggregationNullResponseSumRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new ColumnAggregationNullResponseSumRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<ColumnAggregationNullResponseSumRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required bytes family = 1;
-    public static final int FAMILY_FIELD_NUMBER = 1;
-    private com.google.protobuf.ByteString family_;
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    public boolean hasFamily() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    public com.google.protobuf.ByteString getFamily() {
-      return family_;
-    }
-
-    // optional bytes qualifier = 2;
-    public static final int QUALIFIER_FIELD_NUMBER = 2;
-    private com.google.protobuf.ByteString qualifier_;
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    public boolean hasQualifier() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    public com.google.protobuf.ByteString getQualifier() {
-      return qualifier_;
-    }
-
-    private void initFields() {
-      family_ = com.google.protobuf.ByteString.EMPTY;
-      qualifier_ = com.google.protobuf.ByteString.EMPTY;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasFamily()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, family_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeBytes(2, qualifier_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, family_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(2, qualifier_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest) obj;
-
-      boolean result = true;
-      result = result && (hasFamily() == other.hasFamily());
-      if (hasFamily()) {
-        result = result && getFamily()
-            .equals(other.getFamily());
-      }
-      result = result && (hasQualifier() == other.hasQualifier());
-      if (hasQualifier()) {
-        result = result && getQualifier()
-            .equals(other.getQualifier());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasFamily()) {
-        hash = (37 * hash) + FAMILY_FIELD_NUMBER;
-        hash = (53 * hash) + getFamily().hashCode();
-      }
-      if (hasQualifier()) {
-        hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
-        hash = (53 * hash) + getQualifier().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code ColumnAggregationNullResponseSumRequest}
-     *
-     * <pre>
-     * use unique names for messages in ColumnAggregationXXX.protos due to a bug in
-     * protoc or hadoop's protoc compiler.
-     * </pre>
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        family_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        qualifier_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.family_ = family_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.qualifier_ = qualifier_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.getDefaultInstance()) return this;
-        if (other.hasFamily()) {
-          setFamily(other.getFamily());
-        }
-        if (other.hasQualifier()) {
-          setQualifier(other.getQualifier());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasFamily()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required bytes family = 1;
-      private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public boolean hasFamily() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public com.google.protobuf.ByteString getFamily() {
-        return family_;
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public Builder setFamily(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        family_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public Builder clearFamily() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        family_ = getDefaultInstance().getFamily();
-        onChanged();
-        return this;
-      }
-
-      // optional bytes qualifier = 2;
-      private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public boolean hasQualifier() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public com.google.protobuf.ByteString getQualifier() {
-        return qualifier_;
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public Builder setQualifier(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000002;
-        qualifier_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public Builder clearQualifier() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        qualifier_ = getDefaultInstance().getQualifier();
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:ColumnAggregationNullResponseSumRequest)
-    }
-
-    static {
-      defaultInstance = new ColumnAggregationNullResponseSumRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:ColumnAggregationNullResponseSumRequest)
-  }
-
-  public interface ColumnAggregationNullResponseSumResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // optional int64 sum = 1;
-    /**
-     * <code>optional int64 sum = 1;</code>
-     */
-    boolean hasSum();
-    /**
-     * <code>optional int64 sum = 1;</code>
-     */
-    long getSum();
-  }
-  /**
-   * Protobuf type {@code ColumnAggregationNullResponseSumResponse}
-   */
-  public static final class ColumnAggregationNullResponseSumResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements ColumnAggregationNullResponseSumResponseOrBuilder {
-    // Use ColumnAggregationNullResponseSumResponse.newBuilder() to construct.
-    private ColumnAggregationNullResponseSumResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private ColumnAggregationNullResponseSumResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final ColumnAggregationNullResponseSumResponse defaultInstance;
-    public static ColumnAggregationNullResponseSumResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public ColumnAggregationNullResponseSumResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private ColumnAggregationNullResponseSumResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              sum_ = input.readInt64();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<ColumnAggregationNullResponseSumResponse> PARSER =
-        new com.google.protobuf.AbstractParser<ColumnAggregationNullResponseSumResponse>() {
-      public ColumnAggregationNullResponseSumResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new ColumnAggregationNullResponseSumResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<ColumnAggregationNullResponseSumResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // optional int64 sum = 1;
-    public static final int SUM_FIELD_NUMBER = 1;
-    private long sum_;
-    /**
-     * <code>optional int64 sum = 1;</code>
-     */
-    public boolean hasSum() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>optional int64 sum = 1;</code>
-     */
-    public long getSum() {
-      return sum_;
-    }
-
-    private void initFields() {
-      sum_ = 0L;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeInt64(1, sum_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt64Size(1, sum_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse) obj;
-
-      boolean result = true;
-      result = result && (hasSum() == other.hasSum());
-      if (hasSum()) {
-        result = result && (getSum()
-            == other.getSum());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasSum()) {
-        hash = (37 * hash) + SUM_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getSum());
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code ColumnAggregationNullResponseSumResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        sum_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.internal_static_ColumnAggregationNullResponseSumResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.sum_ = sum_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.getDefaultInstance()) return this;
-        if (other.hasSum()) {
-          setSum(other.getSum());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // optional int64 sum = 1;
-      private long sum_ ;
-      /**
-       * <code>optional int64 sum = 1;</code>
-       */
-      public boolean hasSum() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>optional int64 sum = 1;</code>
-       */
-      public long getSum() {
-        return sum_;
-      }
-      /**
-       * <code>optional int64 sum = 1;</code>
-       */
-      public Builder setSum(long value) {
-        bitField0_ |= 0x00000001;
-        sum_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional int64 sum = 1;</code>
-       */
-      public Builder clearSum() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        sum_ = 0L;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:ColumnAggregationNullResponseSumResponse)
-    }
-
-    static {
-      defaultInstance = new ColumnAggregationNullResponseSumResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:ColumnAggregationNullResponseSumResponse)
-  }
-
-  /**
-   * Protobuf service {@code ColumnAggregationServiceNullResponse}
-   */
-  public static abstract class ColumnAggregationServiceNullResponse
-      implements com.google.protobuf.Service {
-    protected ColumnAggregationServiceNullResponse() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc sum(.ColumnAggregationNullResponseSumRequest) returns (.ColumnAggregationNullResponseSumResponse);</code>
-       */
-      public abstract void sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse> done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new ColumnAggregationServiceNullResponse() {
-        @java.lang.Override
-        public  void sum(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse> done) {
-          impl.sum(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc sum(.ColumnAggregationNullResponseSumRequest) returns (.ColumnAggregationNullResponseSumResponse);</code>
-     */
-    public abstract void sum(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse> done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationServiceNullResponse implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.class,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest request)
-          throws com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithNullResponseProtos.ColumnAggregationNullResponseSumResponse.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:ColumnAggregationServiceNullResponse)
-  }
-
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_ColumnAggregationNullResponseSumRequest_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_ColumnAggregationNullResponseSumRequest_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_ColumnAggregationNullResponseSumResponse_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_ColumnAggregationNullResponseSumResponse_fieldAccessorTable;
-
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n+ColumnAggregationNullResponseProtocol." +
-      "proto\"L\n\'ColumnAggregationNullResponseSu" +
-      "mRequest\022\016\n\006family\030\001 \002(\014\022\021\n\tqualifier\030\002 " +
-      "\001(\014\"7\n(ColumnAggregationNullResponseSumR" +
-      "esponse\022\013\n\003sum\030\001 \001(\0032\202\001\n$ColumnAggregati" +
-      "onServiceNullResponse\022Z\n\003sum\022(.ColumnAgg" +
-      "regationNullResponseSumRequest\032).ColumnA" +
-      "ggregationNullResponseSumResponseBg\n6org" +
-      ".apache.hadoop.hbase.coprocessor.protobu" +
-      "f.generatedB\'ColumnAggregationWithNullRe",
-      "sponseProtos\210\001\001\240\001\001"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_ColumnAggregationNullResponseSumRequest_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_ColumnAggregationNullResponseSumRequest_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_ColumnAggregationNullResponseSumRequest_descriptor,
-              new java.lang.String[] { "Family", "Qualifier", });
-          internal_static_ColumnAggregationNullResponseSumResponse_descriptor =
-            getDescriptor().getMessageTypes().get(1);
-          internal_static_ColumnAggregationNullResponseSumResponse_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_ColumnAggregationNullResponseSumResponse_descriptor,
-              new java.lang.String[] { "Sum", });
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[41/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractMessage.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractMessage.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractMessage.java
deleted file mode 100644
index 4a6fefa..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractMessage.java
+++ /dev/null
@@ -1,646 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLite;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.util.Arrays;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-
-/**
- * A partial implementation of the {@link Message} interface which implements
- * as many methods of that interface as possible in terms of other methods.
- *
- * @author kenton@google.com Kenton Varda
- */
-public abstract class AbstractMessage
-    // TODO(dweis): Update GeneratedMessage to parameterize with MessageType and BuilderType.
-    extends AbstractMessageLite
-    implements Message {
-
-  @Override
-  public boolean isInitialized() {
-    return MessageReflection.isInitialized(this);
-  }
-
-  /**
-   * Interface for the parent of a Builder that allows the builder to
-   * communicate invalidations back to the parent for use when using nested
-   * builders.
-   */
-  protected interface BuilderParent {
-
-    /**
-     * A builder becomes dirty whenever a field is modified -- including fields
-     * in nested builders -- and becomes clean when build() is called.  Thus,
-     * when a builder becomes dirty, all its parents become dirty as well, and
-     * when it becomes clean, all its children become clean.  The dirtiness
-     * state is used to invalidate certain cached values.
-     * <br>
-     * To this end, a builder calls markDirty() on its parent whenever it
-     * transitions from clean to dirty.  The parent must propagate this call to
-     * its own parent, unless it was already dirty, in which case the
-     * grandparent must necessarily already be dirty as well.  The parent can
-     * only transition back to "clean" after calling build() on all children.
-     */
-    void markDirty();
-  }
-
-  /** Create a nested builder. */
-  protected Message.Builder newBuilderForType(BuilderParent parent) {
-    throw new UnsupportedOperationException("Nested builder is not supported for this type.");
-  }
-
-
-  @Override
-  public List<String> findInitializationErrors() {
-    return MessageReflection.findMissingFields(this);
-  }
-
-  @Override
-  public String getInitializationErrorString() {
-    return MessageReflection.delimitWithCommas(findInitializationErrors());
-  }
-
-  /** TODO(jieluo): Clear it when all subclasses have implemented this method. */
-  @Override
-  public boolean hasOneof(OneofDescriptor oneof) {
-    throw new UnsupportedOperationException("hasOneof() is not implemented.");
-  }
-
-  /** TODO(jieluo): Clear it when all subclasses have implemented this method. */
-  @Override
-  public FieldDescriptor getOneofFieldDescriptor(OneofDescriptor oneof) {
-    throw new UnsupportedOperationException(
-        "getOneofFieldDescriptor() is not implemented.");
-  }
-
-  @Override
-  public final String toString() {
-    return TextFormat.printToString(this);
-  }
-
-  @Override
-  public void writeTo(final CodedOutputStream output) throws IOException {
-    MessageReflection.writeMessageTo(this, getAllFields(), output, false);
-  }
-
-  protected int memoizedSize = -1;
-
-  @Override
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) {
-      return size;
-    }
-
-    memoizedSize = MessageReflection.getSerializedSize(this, getAllFields());
-    return memoizedSize;
-  }
-
-  @Override
-  public boolean equals(final Object other) {
-    if (other == this) {
-      return true;
-    }
-    if (!(other instanceof Message)) {
-      return false;
-    }
-    final Message otherMessage = (Message) other;
-    if (getDescriptorForType() != otherMessage.getDescriptorForType()) {
-      return false;
-    }
-    return compareFields(getAllFields(), otherMessage.getAllFields()) &&
-        getUnknownFields().equals(otherMessage.getUnknownFields());
-  }
-
-  @Override
-  public int hashCode() {
-    int hash = memoizedHashCode;
-    if (hash == 0) {
-      hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      hash = hashFields(hash, getAllFields());
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-    }
-    return hash;
-  }
-  
-  private static ByteString toByteString(Object value) {
-    if (value instanceof byte[]) {
-      return ByteString.copyFrom((byte[]) value);
-    } else {
-      return (ByteString) value;
-    }
-  }
- 
-  /**
-   * Compares two bytes fields. The parameters must be either a byte array or a
-   * ByteString object. They can be of different type though.
-   */
-  private static boolean compareBytes(Object a, Object b) {
-    if (a instanceof byte[] && b instanceof byte[]) {
-      return Arrays.equals((byte[])a, (byte[])b);
-    }
-    return toByteString(a).equals(toByteString(b));
-  }
-  
-  /**
-   * Converts a list of MapEntry messages into a Map used for equals() and
-   * hashCode().
-   */
-  @SuppressWarnings({"rawtypes", "unchecked"})
-  private static Map convertMapEntryListToMap(List list) {
-    if (list.isEmpty()) {
-      return Collections.emptyMap();
-    }
-    Map result = new HashMap();
-    Iterator iterator = list.iterator();
-    Message entry = (Message) iterator.next();
-    Descriptors.Descriptor descriptor = entry.getDescriptorForType();
-    Descriptors.FieldDescriptor key = descriptor.findFieldByName("key");
-    Descriptors.FieldDescriptor value = descriptor.findFieldByName("value");
-    Object fieldValue = entry.getField(value);
-    if (fieldValue instanceof EnumValueDescriptor) {
-      fieldValue = ((EnumValueDescriptor) fieldValue).getNumber();
-    }
-    result.put(entry.getField(key), fieldValue);
-    while (iterator.hasNext()) {
-      entry = (Message) iterator.next();
-      fieldValue = entry.getField(value);
-      if (fieldValue instanceof EnumValueDescriptor) {
-        fieldValue = ((EnumValueDescriptor) fieldValue).getNumber();
-      }
-      result.put(entry.getField(key), fieldValue);
-    }
-    return result;
-  }
-  
-  /**
-   * Compares two map fields. The parameters must be a list of MapEntry
-   * messages.
-   */
-  @SuppressWarnings({"rawtypes", "unchecked"})
-  private static boolean compareMapField(Object a, Object b) {
-    Map ma = convertMapEntryListToMap((List) a);
-    Map mb = convertMapEntryListToMap((List) b);
-    return MapFieldLite.equals(ma, mb);
-  }
-  
-  /**
-   * Compares two set of fields.
-   * This method is used to implement {@link AbstractMessage#equals(Object)}
-   * and {@link AbstractMutableMessage#equals(Object)}. It takes special care
-   * of bytes fields because immutable messages and mutable messages use
-   * different Java type to reprensent a bytes field and this method should be
-   * able to compare immutable messages, mutable messages and also an immutable
-   * message to a mutable message.
-   */
-  static boolean compareFields(Map<FieldDescriptor, Object> a,
-      Map<FieldDescriptor, Object> b) {
-    if (a.size() != b.size()) {
-      return false;
-    }
-    for (FieldDescriptor descriptor : a.keySet()) {
-      if (!b.containsKey(descriptor)) {
-        return false;
-      }
-      Object value1 = a.get(descriptor);
-      Object value2 = b.get(descriptor);
-      if (descriptor.getType() == FieldDescriptor.Type.BYTES) {
-        if (descriptor.isRepeated()) {
-          List list1 = (List) value1;
-          List list2 = (List) value2;
-          if (list1.size() != list2.size()) {
-            return false;
-          }
-          for (int i = 0; i < list1.size(); i++) {
-            if (!compareBytes(list1.get(i), list2.get(i))) {
-              return false;
-            }
-          }
-        } else {
-          // Compares a singular bytes field.
-          if (!compareBytes(value1, value2)) {
-            return false;
-          }
-        }
-      } else if (descriptor.isMapField()) {
-        if (!compareMapField(value1, value2)) {
-          return false;
-        }
-      } else {
-        // Compare non-bytes fields.
-        if (!value1.equals(value2)) {
-          return false;
-        }
-      }
-    }
-    return true;
-  }
-  
-  /**
-   * Calculates the hash code of a map field. {@code value} must be a list of
-   * MapEntry messages.
-   */
-  @SuppressWarnings("unchecked")
-  private static int hashMapField(Object value) {
-    return MapFieldLite.calculateHashCodeForMap(convertMapEntryListToMap((List) value));
-  }
-
-  /** Get a hash code for given fields and values, using the given seed. */
-  @SuppressWarnings("unchecked")
-  protected static int hashFields(int hash, Map<FieldDescriptor, Object> map) {
-    for (Map.Entry<FieldDescriptor, Object> entry : map.entrySet()) {
-      FieldDescriptor field = entry.getKey();
-      Object value = entry.getValue();
-      hash = (37 * hash) + field.getNumber();
-      if (field.isMapField()) {
-        hash = (53 * hash) + hashMapField(value);
-      } else if (field.getType() != FieldDescriptor.Type.ENUM){
-        hash = (53 * hash) + value.hashCode();
-      } else if (field.isRepeated()) {
-        List<? extends EnumLite> list = (List<? extends EnumLite>) value;
-        hash = (53 * hash) + Internal.hashEnumList(list);
-      } else {
-        hash = (53 * hash) + Internal.hashEnum((EnumLite) value);
-      }
-    }
-    return hash;
-  }
-
-  /**
-   * Package private helper method for AbstractParser to create
-   * UninitializedMessageException with missing field information.
-   */
-  @Override
-  UninitializedMessageException newUninitializedMessageException() {
-    return Builder.newUninitializedMessageException(this);
-  }
-
-  // =================================================================
-
-  /**
-   * A partial implementation of the {@link Message.Builder} interface which
-   * implements as many methods of that interface as possible in terms of
-   * other methods.
-   */
-  @SuppressWarnings("unchecked")
-  public static abstract class Builder<BuilderType extends Builder<BuilderType>>
-      extends AbstractMessageLite.Builder
-      implements Message.Builder {
-    // The compiler produces an error if this is not declared explicitly.
-    @Override
-    public abstract BuilderType clone();
-
-    /** TODO(jieluo): Clear it when all subclasses have implemented this method. */
-    @Override
-    public boolean hasOneof(OneofDescriptor oneof) {
-      throw new UnsupportedOperationException("hasOneof() is not implemented.");
-    }
-
-    /** TODO(jieluo): Clear it when all subclasses have implemented this method. */
-    @Override
-    public FieldDescriptor getOneofFieldDescriptor(OneofDescriptor oneof) {
-      throw new UnsupportedOperationException(
-          "getOneofFieldDescriptor() is not implemented.");
-    }
-
-    /** TODO(jieluo): Clear it when all subclasses have implemented this method. */
-    @Override
-    public BuilderType clearOneof(OneofDescriptor oneof) {
-      throw new UnsupportedOperationException("clearOneof() is not implemented.");
-    }
-
-    @Override
-    public BuilderType clear() {
-      for (final Map.Entry<FieldDescriptor, Object> entry :
-           getAllFields().entrySet()) {
-        clearField(entry.getKey());
-      }
-      return (BuilderType) this;
-    }
-
-    @Override
-    public List<String> findInitializationErrors() {
-      return MessageReflection.findMissingFields(this);
-    }
-
-    @Override
-    public String getInitializationErrorString() {
-      return MessageReflection.delimitWithCommas(findInitializationErrors());
-    }
-    
-    @Override
-    protected BuilderType internalMergeFrom(AbstractMessageLite other) {
-      return mergeFrom((Message) other);
-    }
-
-    @Override
-    public BuilderType mergeFrom(final Message other) {
-      if (other.getDescriptorForType() != getDescriptorForType()) {
-        throw new IllegalArgumentException(
-          "mergeFrom(Message) can only merge messages of the same type.");
-      }
-
-      // Note:  We don't attempt to verify that other's fields have valid
-      //   types.  Doing so would be a losing battle.  We'd have to verify
-      //   all sub-messages as well, and we'd have to make copies of all of
-      //   them to insure that they don't change after verification (since
-      //   the Message interface itself cannot enforce immutability of
-      //   implementations).
-      // TODO(kenton):  Provide a function somewhere called makeDeepCopy()
-      //   which allows people to make secure deep copies of messages.
-
-      for (final Map.Entry<FieldDescriptor, Object> entry :
-           other.getAllFields().entrySet()) {
-        final FieldDescriptor field = entry.getKey();
-        if (field.isRepeated()) {
-          for (final Object element : (List)entry.getValue()) {
-            addRepeatedField(field, element);
-          }
-        } else if (field.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-          final Message existingValue = (Message)getField(field);
-          if (existingValue == existingValue.getDefaultInstanceForType()) {
-            setField(field, entry.getValue());
-          } else {
-            setField(field,
-              existingValue.newBuilderForType()
-                .mergeFrom(existingValue)
-                .mergeFrom((Message)entry.getValue())
-                .build());
-          }
-        } else {
-          setField(field, entry.getValue());
-        }
-      }
-
-      mergeUnknownFields(other.getUnknownFields());
-
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType mergeFrom(final CodedInputStream input)
-                                 throws IOException {
-      return mergeFrom(input, ExtensionRegistry.getEmptyRegistry());
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final CodedInputStream input,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      final UnknownFieldSet.Builder unknownFields =
-        UnknownFieldSet.newBuilder(getUnknownFields());
-      while (true) {
-        final int tag = input.readTag();
-        if (tag == 0) {
-          break;
-        }
-
-        MessageReflection.BuilderAdapter builderAdapter =
-            new MessageReflection.BuilderAdapter(this);
-        if (!MessageReflection.mergeFieldFrom(input, unknownFields,
-                                              extensionRegistry,
-                                              getDescriptorForType(),
-                                              builderAdapter,
-                                              tag)) {
-          // end group tag
-          break;
-        }
-      }
-      setUnknownFields(unknownFields.build());
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType mergeUnknownFields(final UnknownFieldSet unknownFields) {
-      setUnknownFields(
-        UnknownFieldSet.newBuilder(getUnknownFields())
-                       .mergeFrom(unknownFields)
-                       .build());
-      return (BuilderType) this;
-    }
-
-    @Override
-    public Message.Builder getFieldBuilder(final FieldDescriptor field) {
-      throw new UnsupportedOperationException(
-          "getFieldBuilder() called on an unsupported message type.");
-    }
-
-    @Override
-    public Message.Builder getRepeatedFieldBuilder(final FieldDescriptor field, int index) {
-      throw new UnsupportedOperationException(
-          "getRepeatedFieldBuilder() called on an unsupported message type.");
-    }
-
-    @Override
-    public String toString() {
-      return TextFormat.printToString(this);
-    }
-
-    /**
-     * Construct an UninitializedMessageException reporting missing fields in
-     * the given message.
-     */
-    protected static UninitializedMessageException
-        newUninitializedMessageException(Message message) {
-      return new UninitializedMessageException(
-          MessageReflection.findMissingFields(message));
-    }
-
-    /**
-     * Used to support nested builders and called to mark this builder as clean.
-     * Clean builders will propagate the {@link BuilderParent#markDirty()} event
-     * to their parent builders, while dirty builders will not, as their parents
-     * should be dirty already.
-     *
-     * NOTE: Implementations that don't support nested builders don't need to
-     * override this method.
-     */
-    void markClean() {
-      throw new IllegalStateException("Should be overridden by subclasses.");
-    }
-
-    /**
-     * Used to support nested builders and called when this nested builder is
-     * no longer used by its parent builder and should release the reference
-     * to its parent builder.
-     *
-     * NOTE: Implementations that don't support nested builders don't need to
-     * override this method.
-     */
-    void dispose() {
-      throw new IllegalStateException("Should be overridden by subclasses.");
-    }
-
-    // ===============================================================
-    // The following definitions seem to be required in order to make javac
-    // not produce weird errors like:
-    //
-    // java/org.apache.hadoop.hbase.shaded.com.google.protobuf/DynamicMessage.java:203: types
-    //   org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessage.Builder<
-    //     org.apache.hadoop.hbase.shaded.com.google.protobuf.DynamicMessage.Builder> and
-    //   org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessage.Builder<
-    //     org.apache.hadoop.hbase.shaded.com.google.protobuf.DynamicMessage.Builder> are incompatible; both
-    //   define mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString), but with unrelated
-    //   return types.
-    //
-    // Strangely, these lines are only needed if javac is invoked separately
-    // on AbstractMessage.java and AbstractMessageLite.java.  If javac is
-    // invoked on both simultaneously, it works.  (Or maybe the important
-    // point is whether or not DynamicMessage.java is compiled together with
-    // AbstractMessageLite.java -- not sure.)  I suspect this is a compiler
-    // bug.
-
-    @Override
-    public BuilderType mergeFrom(final ByteString data)
-        throws InvalidProtocolBufferException {
-      return (BuilderType) super.mergeFrom(data);
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final ByteString data,
-        final ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      return (BuilderType) super.mergeFrom(data, extensionRegistry);
-    }
-
-    @Override
-    public BuilderType mergeFrom(final byte[] data)
-        throws InvalidProtocolBufferException {
-      return (BuilderType) super.mergeFrom(data);
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final byte[] data, final int off, final int len)
-        throws InvalidProtocolBufferException {
-      return (BuilderType) super.mergeFrom(data, off, len);
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final byte[] data,
-        final ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      return (BuilderType) super.mergeFrom(data, extensionRegistry);
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final byte[] data, final int off, final int len,
-        final ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      return (BuilderType) super.mergeFrom(data, off, len, extensionRegistry);
-    }
-
-    @Override
-    public BuilderType mergeFrom(final InputStream input)
-        throws IOException {
-      return (BuilderType) super.mergeFrom(input);
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final InputStream input,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      return (BuilderType) super.mergeFrom(input, extensionRegistry);
-    }
-
-    @Override
-    public boolean mergeDelimitedFrom(final InputStream input)
-        throws IOException {
-      return super.mergeDelimitedFrom(input);
-    }
-
-    @Override
-    public boolean mergeDelimitedFrom(
-        final InputStream input,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      return super.mergeDelimitedFrom(input, extensionRegistry);
-    }
-  }
-
-  /**
-   * @deprecated from v3.0.0-beta-3+, for compatibility with v2.5.0 and v2.6.1
-   * generated code.
-   */
-  @Deprecated
-  protected static int hashLong(long n) {
-    return (int) (n ^ (n >>> 32));
-  }
-  //
-  /**
-   * @deprecated from v3.0.0-beta-3+, for compatibility with v2.5.0 and v2.6.1
-   * generated code.
-   */
-  @Deprecated
-  protected static int hashBoolean(boolean b) {
-    return b ? 1231 : 1237;
-  }
-  //
-  /**
-   * @deprecated from v3.0.0-beta-3+, for compatibility with v2.5.0 and v2.6.1
-   * generated code.
-   */
-  @Deprecated
-  protected static int hashEnum(EnumLite e) {
-    return e.getNumber();
-  }
-  //
-  /**
-   * @deprecated from v3.0.0-beta-3+, for compatibility with v2.5.0 and v2.6.1
-   * generated code.
-   */
-  @Deprecated
-  protected static int hashEnumList(List<? extends EnumLite> list) {
-    int hash = 1;
-    for (EnumLite e : list) {
-      hash = 31 * hash + hashEnum(e);
-    }
-    return hash;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractMessageLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractMessageLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractMessageLite.java
deleted file mode 100644
index e5b87f0..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractMessageLite.java
+++ /dev/null
@@ -1,383 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.FilterInputStream;
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.OutputStream;
-import java.util.Collection;
-
-/**
- * A partial implementation of the {@link MessageLite} interface which
- * implements as many methods of that interface as possible in terms of other
- * methods.
- *
- * @author kenton@google.com Kenton Varda
- */
-public abstract class AbstractMessageLite<
-    MessageType extends AbstractMessageLite<MessageType, BuilderType>,
-    BuilderType extends AbstractMessageLite.Builder<MessageType, BuilderType>>
-        implements MessageLite {
-  protected int memoizedHashCode = 0;
-  @Override
-  public ByteString toByteString() {
-    try {
-      final ByteString.CodedBuilder out =
-        ByteString.newCodedBuilder(getSerializedSize());
-      writeTo(out.getCodedOutput());
-      return out.build();
-    } catch (IOException e) {
-      throw new RuntimeException(getSerializingExceptionMessage("ByteString"), e);
-    }
-  }
-
-  @Override
-  public byte[] toByteArray() {
-    try {
-      final byte[] result = new byte[getSerializedSize()];
-      final CodedOutputStream output = CodedOutputStream.newInstance(result);
-      writeTo(output);
-      output.checkNoSpaceLeft();
-      return result;
-    } catch (IOException e) {
-      throw new RuntimeException(getSerializingExceptionMessage("byte array"), e);
-    }
-  }
-
-  @Override
-  public void writeTo(final OutputStream output) throws IOException {
-    final int bufferSize =
-        CodedOutputStream.computePreferredBufferSize(getSerializedSize());
-    final CodedOutputStream codedOutput =
-        CodedOutputStream.newInstance(output, bufferSize);
-    writeTo(codedOutput);
-    codedOutput.flush();
-  }
-
-  @Override
-  public void writeDelimitedTo(final OutputStream output) throws IOException {
-    final int serialized = getSerializedSize();
-    final int bufferSize = CodedOutputStream.computePreferredBufferSize(
-        CodedOutputStream.computeRawVarint32Size(serialized) + serialized);
-    final CodedOutputStream codedOutput =
-        CodedOutputStream.newInstance(output, bufferSize);
-    codedOutput.writeRawVarint32(serialized);
-    writeTo(codedOutput);
-    codedOutput.flush();
-  }
-
-
-  /**
-   * Package private helper method for AbstractParser to create
-   * UninitializedMessageException.
-   */
-  UninitializedMessageException newUninitializedMessageException() {
-    return new UninitializedMessageException(this);
-  }
-
-  private String getSerializingExceptionMessage(String target) {
-    return "Serializing " + getClass().getName() + " to a " + target
-        + " threw an IOException (should never happen).";
-  }
-
-  protected static void checkByteStringIsUtf8(ByteString byteString)
-      throws IllegalArgumentException {
-    if (!byteString.isValidUtf8()) {
-      throw new IllegalArgumentException("Byte string is not UTF-8.");
-    }
-  }
-
-  protected static <T> void addAll(final Iterable<T> values,
-      final Collection<? super T> list) {
-    Builder.addAll(values, list);
-  }
-
-  /**
-   * A partial implementation of the {@link Message.Builder} interface which
-   * implements as many methods of that interface as possible in terms of
-   * other methods.
-   */
-  @SuppressWarnings("unchecked")
-  public abstract static class Builder<
-      MessageType extends AbstractMessageLite<MessageType, BuilderType>,
-      BuilderType extends Builder<MessageType, BuilderType>>
-      implements MessageLite.Builder {
-    // The compiler produces an error if this is not declared explicitly.
-    @Override
-    public abstract BuilderType clone();
-
-    @Override
-    public BuilderType mergeFrom(final CodedInputStream input) throws IOException {
-      return mergeFrom(input, ExtensionRegistryLite.getEmptyRegistry());
-    }
-
-    // Re-defined here for return type covariance.
-    @Override
-    public abstract BuilderType mergeFrom(
-        final CodedInputStream input, final ExtensionRegistryLite extensionRegistry)
-        throws IOException;
-
-    @Override
-    public BuilderType mergeFrom(final ByteString data) throws InvalidProtocolBufferException {
-      try {
-        final CodedInputStream input = data.newCodedInput();
-        mergeFrom(input);
-        input.checkLastTagWas(0);
-        return (BuilderType) this;
-      } catch (InvalidProtocolBufferException e) {
-        throw e;
-      } catch (IOException e) {
-        throw new RuntimeException(getReadingExceptionMessage("ByteString"), e);
-      }
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final ByteString data, final ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      try {
-        final CodedInputStream input = data.newCodedInput();
-        mergeFrom(input, extensionRegistry);
-        input.checkLastTagWas(0);
-        return (BuilderType) this;
-      } catch (InvalidProtocolBufferException e) {
-        throw e;
-      } catch (IOException e) {
-        throw new RuntimeException(getReadingExceptionMessage("ByteString"), e);
-      }
-    }
-
-    @Override
-    public BuilderType mergeFrom(final byte[] data) throws InvalidProtocolBufferException {
-      return mergeFrom(data, 0, data.length);
-    }
-
-    @Override
-    public BuilderType mergeFrom(final byte[] data, final int off, final int len)
-        throws InvalidProtocolBufferException {
-      try {
-        final CodedInputStream input =
-            CodedInputStream.newInstance(data, off, len);
-        mergeFrom(input);
-        input.checkLastTagWas(0);
-        return (BuilderType) this;
-      } catch (InvalidProtocolBufferException e) {
-        throw e;
-      } catch (IOException e) {
-        throw new RuntimeException(getReadingExceptionMessage("byte array"), e);
-      }
-    }
-
-    @Override
-    public BuilderType mergeFrom(final byte[] data, final ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      return mergeFrom(data, 0, data.length, extensionRegistry);
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final byte[] data,
-        final int off,
-        final int len,
-        final ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      try {
-        final CodedInputStream input =
-            CodedInputStream.newInstance(data, off, len);
-        mergeFrom(input, extensionRegistry);
-        input.checkLastTagWas(0);
-        return (BuilderType) this;
-      } catch (InvalidProtocolBufferException e) {
-        throw e;
-      } catch (IOException e) {
-        throw new RuntimeException(getReadingExceptionMessage("byte array"), e);
-      }
-    }
-
-    @Override
-    public BuilderType mergeFrom(final InputStream input) throws IOException {
-      final CodedInputStream codedInput = CodedInputStream.newInstance(input);
-      mergeFrom(codedInput);
-      codedInput.checkLastTagWas(0);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        final InputStream input, final ExtensionRegistryLite extensionRegistry) throws IOException {
-      final CodedInputStream codedInput = CodedInputStream.newInstance(input);
-      mergeFrom(codedInput, extensionRegistry);
-      codedInput.checkLastTagWas(0);
-      return (BuilderType) this;
-    }
-
-    /**
-     * An InputStream implementations which reads from some other InputStream
-     * but is limited to a particular number of bytes.  Used by
-     * mergeDelimitedFrom().  This is intentionally package-private so that
-     * UnknownFieldSet can share it.
-     */
-    static final class LimitedInputStream extends FilterInputStream {
-      private int limit;
-
-      LimitedInputStream(InputStream in, int limit) {
-        super(in);
-        this.limit = limit;
-      }
-
-      @Override
-      public int available() throws IOException {
-        return Math.min(super.available(), limit);
-      }
-
-      @Override
-      public int read() throws IOException {
-        if (limit <= 0) {
-          return -1;
-        }
-        final int result = super.read();
-        if (result >= 0) {
-          --limit;
-        }
-        return result;
-      }
-
-      @Override
-      public int read(final byte[] b, final int off, int len)
-                      throws IOException {
-        if (limit <= 0) {
-          return -1;
-        }
-        len = Math.min(len, limit);
-        final int result = super.read(b, off, len);
-        if (result >= 0) {
-          limit -= result;
-        }
-        return result;
-      }
-
-      @Override
-      public long skip(final long n) throws IOException {
-        final long result = super.skip(Math.min(n, limit));
-        if (result >= 0) {
-          limit -= result;
-        }
-        return result;
-      }
-    }
-
-    @Override
-    public boolean mergeDelimitedFrom(
-        final InputStream input, final ExtensionRegistryLite extensionRegistry) throws IOException {
-      final int firstByte = input.read();
-      if (firstByte == -1) {
-        return false;
-      }
-      final int size = CodedInputStream.readRawVarint32(firstByte, input);
-      final InputStream limitedInput = new LimitedInputStream(input, size);
-      mergeFrom(limitedInput, extensionRegistry);
-      return true;
-    }
-
-    @Override
-    public boolean mergeDelimitedFrom(final InputStream input) throws IOException {
-      return mergeDelimitedFrom(input,
-          ExtensionRegistryLite.getEmptyRegistry());
-    }
-
-    @Override
-    @SuppressWarnings("unchecked") // isInstance takes care of this
-    public BuilderType mergeFrom(final MessageLite other) {
-      if (!getDefaultInstanceForType().getClass().isInstance(other)) {
-        throw new IllegalArgumentException(
-            "mergeFrom(MessageLite) can only merge messages of the same type.");
-      }
-
-      return internalMergeFrom((MessageType) other);
-    }
-
-    protected abstract BuilderType internalMergeFrom(MessageType message);
-
-    private String getReadingExceptionMessage(String target) {
-      return "Reading " + getClass().getName() + " from a " + target
-          + " threw an IOException (should never happen).";
-    }
-
-    /**
-     * Construct an UninitializedMessageException reporting missing fields in
-     * the given message.
-     */
-    protected static UninitializedMessageException
-        newUninitializedMessageException(MessageLite message) {
-      return new UninitializedMessageException(message);
-    }
-
-    /**
-     * Adds the {@code values} to the {@code list}.  This is a helper method
-     * used by generated code.  Users should ignore it.
-     *
-     * @throws NullPointerException if {@code values} or any of the elements of
-     * {@code values} is null. When that happens, some elements of
-     * {@code values} may have already been added to the result {@code list}.
-     */
-    protected static <T> void addAll(final Iterable<T> values,
-                                     final Collection<? super T> list) {
-      if (values == null) {
-        throw new NullPointerException();
-      }
-      if (values instanceof LazyStringList) {
-        // For StringOrByteStringLists, check the underlying elements to avoid
-        // forcing conversions of ByteStrings to Strings.
-        checkForNullValues(((LazyStringList) values).getUnderlyingElements());
-        list.addAll((Collection<T>) values);
-      } else if (values instanceof Collection) {
-        checkForNullValues(values);
-        list.addAll((Collection<T>) values);
-      } else {
-        for (final T value : values) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          list.add(value);
-        }
-      }
-    }
-
-    private static void checkForNullValues(final Iterable<?> values) {
-      for (final Object value : values) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-      }
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractParser.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractParser.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractParser.java
deleted file mode 100644
index 5fe6a22..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractParser.java
+++ /dev/null
@@ -1,258 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.LimitedInputStream;
-
-import java.io.IOException;
-import java.io.InputStream;
-
-/**
- * A partial implementation of the {@link Parser} interface which implements
- * as many methods of that interface as possible in terms of other methods.
- *
- * Note: This class implements all the convenience methods in the
- * {@link Parser} interface. See {@link Parser} for related javadocs.
- * Subclasses need to implement
- * {@link Parser#parsePartialFrom(CodedInputStream, ExtensionRegistryLite)}
- *
- * @author liujisi@google.com (Pherl Liu)
- */
-public abstract class AbstractParser<MessageType extends MessageLite>
-    implements Parser<MessageType> {
-  /**
-   * Creates an UninitializedMessageException for MessageType.
-   */
-  private UninitializedMessageException
-      newUninitializedMessageException(MessageType message) {
-    if (message instanceof AbstractMessageLite) {
-      return ((AbstractMessageLite) message).newUninitializedMessageException();
-    }
-    return new UninitializedMessageException(message);
-  }
-
-  /**
-   * Helper method to check if message is initialized.
-   *
-   * @throws InvalidProtocolBufferException if it is not initialized.
-   * @return The message to check.
-   */
-  private MessageType checkMessageInitialized(MessageType message)
-      throws InvalidProtocolBufferException {
-    if (message != null && !message.isInitialized()) {
-      throw newUninitializedMessageException(message)
-          .asInvalidProtocolBufferException()
-          .setUnfinishedMessage(message);
-    }
-    return message;
-  }
-
-  private static final ExtensionRegistryLite EMPTY_REGISTRY
-      = ExtensionRegistryLite.getEmptyRegistry();
-
-  @Override
-  public MessageType parsePartialFrom(CodedInputStream input)
-      throws InvalidProtocolBufferException {
-    return parsePartialFrom(input, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parseFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialFrom(input, extensionRegistry));
-  }
-
-  @Override
-  public MessageType parseFrom(CodedInputStream input) throws InvalidProtocolBufferException {
-    return parseFrom(input, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parsePartialFrom(ByteString data, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    MessageType message;
-    try {
-      CodedInputStream input = data.newCodedInput();
-      message = parsePartialFrom(input, extensionRegistry);
-      try {
-        input.checkLastTagWas(0);
-      } catch (InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(message);
-      }
-      return message;
-    } catch (InvalidProtocolBufferException e) {
-      throw e;
-    }
-  }
-
-  @Override
-  public MessageType parsePartialFrom(ByteString data) throws InvalidProtocolBufferException {
-    return parsePartialFrom(data, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parseFrom(ByteString data, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(parsePartialFrom(data, extensionRegistry));
-  }
-
-  @Override
-  public MessageType parseFrom(ByteString data) throws InvalidProtocolBufferException {
-    return parseFrom(data, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parsePartialFrom(
-      byte[] data, int off, int len, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    try {
-      CodedInputStream input = CodedInputStream.newInstance(data, off, len);
-      MessageType message = parsePartialFrom(input, extensionRegistry);
-      try {
-        input.checkLastTagWas(0);
-      } catch (InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(message);
-      }
-      return message;
-    } catch (InvalidProtocolBufferException e) {
-      throw e;
-    }
-  }
-
-  @Override
-  public MessageType parsePartialFrom(byte[] data, int off, int len)
-      throws InvalidProtocolBufferException {
-    return parsePartialFrom(data, off, len, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parsePartialFrom(byte[] data, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return parsePartialFrom(data, 0, data.length, extensionRegistry);
-  }
-
-  @Override
-  public MessageType parsePartialFrom(byte[] data) throws InvalidProtocolBufferException {
-    return parsePartialFrom(data, 0, data.length, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parseFrom(
-      byte[] data, int off, int len, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialFrom(data, off, len, extensionRegistry));
-  }
-
-  @Override
-  public MessageType parseFrom(byte[] data, int off, int len)
-      throws InvalidProtocolBufferException {
-    return parseFrom(data, off, len, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parseFrom(byte[] data, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return parseFrom(data, 0, data.length, extensionRegistry);
-  }
-
-  @Override
-  public MessageType parseFrom(byte[] data) throws InvalidProtocolBufferException {
-    return parseFrom(data, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parsePartialFrom(InputStream input, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    CodedInputStream codedInput = CodedInputStream.newInstance(input);
-    MessageType message = parsePartialFrom(codedInput, extensionRegistry);
-    try {
-      codedInput.checkLastTagWas(0);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(message);
-    }
-    return message;
-  }
-
-  @Override
-  public MessageType parsePartialFrom(InputStream input) throws InvalidProtocolBufferException {
-    return parsePartialFrom(input, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parseFrom(InputStream input, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialFrom(input, extensionRegistry));
-  }
-
-  @Override
-  public MessageType parseFrom(InputStream input) throws InvalidProtocolBufferException {
-    return parseFrom(input, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parsePartialDelimitedFrom(
-      InputStream input, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    int size;
-    try {
-      int firstByte = input.read();
-      if (firstByte == -1) {
-        return null;
-      }
-      size = CodedInputStream.readRawVarint32(firstByte, input);
-    } catch (IOException e) {
-      throw new InvalidProtocolBufferException(e);
-    }
-    InputStream limitedInput = new LimitedInputStream(input, size);
-    return parsePartialFrom(limitedInput, extensionRegistry);
-  }
-
-  @Override
-  public MessageType parsePartialDelimitedFrom(InputStream input)
-      throws InvalidProtocolBufferException {
-    return parsePartialDelimitedFrom(input, EMPTY_REGISTRY);
-  }
-
-  @Override
-  public MessageType parseDelimitedFrom(InputStream input, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialDelimitedFrom(input, extensionRegistry));
-  }
-
-  @Override
-  public MessageType parseDelimitedFrom(InputStream input) throws InvalidProtocolBufferException {
-    return parseDelimitedFrom(input, EMPTY_REGISTRY);
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractProtobufList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractProtobufList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractProtobufList.java
deleted file mode 100644
index 87b0cdc..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AbstractProtobufList.java
+++ /dev/null
@@ -1,180 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.ProtobufList;
-
-import java.util.AbstractList;
-import java.util.Collection;
-import java.util.List;
-import java.util.RandomAccess;
-
-/**
- * An abstract implementation of {@link ProtobufList} which manages mutability semantics. All mutate
- * methods must check if the list is mutable before proceeding. Subclasses must invoke
- * {@link #ensureIsMutable()} manually when overriding those methods.
- * <p>
- * This implementation assumes all subclasses are array based, supporting random access.
- */
-abstract class AbstractProtobufList<E> extends AbstractList<E> implements ProtobufList<E> {
-
-  protected static final int DEFAULT_CAPACITY = 10;
-
-  /**
-   * Whether or not this list is modifiable.
-   */
-  private boolean isMutable;
-
-  /**
-   * Constructs a mutable list by default.
-   */
-  AbstractProtobufList() {
-    isMutable = true;
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (o == this) {
-      return true;
-    }
-    if (!(o instanceof List)) {
-      return false;
-    }
-    // Handle lists that do not support RandomAccess as efficiently as possible by using an iterator
-    // based approach in our super class. Otherwise our index based approach will avoid those
-    // allocations.
-    if (!(o instanceof RandomAccess)) {
-      return super.equals(o);
-    }
-
-    List<?> other = (List<?>) o;
-    final int size = size();
-    if (size != other.size()) {
-      return false;
-    }
-    for (int i = 0; i < size; i++) {
-      if (!get(i).equals(other.get(i))) {
-        return false;
-      }
-    }
-    return true;
-  }
-
-  @Override
-  public int hashCode() {
-    final int size = size();
-    int hashCode = 1;
-    for (int i = 0; i < size; i++) {
-      hashCode = (31 * hashCode) + get(i).hashCode();
-    }
-    return hashCode;
-  }
-
-  @Override
-  public boolean add(E e) {
-    ensureIsMutable();
-    return super.add(e);
-  }
-
-  @Override
-  public void add(int index, E element) {
-    ensureIsMutable();
-    super.add(index, element);
-  }
-
-  @Override
-  public boolean addAll(Collection<? extends E> c) {
-    ensureIsMutable();
-    return super.addAll(c);
-  }
-  
-  @Override
-  public boolean addAll(int index, Collection<? extends E> c) {
-    ensureIsMutable();
-    return super.addAll(index, c);
-  }
-
-  @Override
-  public void clear() {
-    ensureIsMutable();
-    super.clear();
-  }
-  
-  @Override
-  public boolean isModifiable() {
-    return isMutable;
-  }
-  
-  @Override
-  public final void makeImmutable() {
-    isMutable = false;
-  }
-  
-  @Override
-  public E remove(int index) {
-    ensureIsMutable();
-    return super.remove(index);
-  }
-  
-  @Override
-  public boolean remove(Object o) {
-    ensureIsMutable();
-    return super.remove(o);
-  }
-  
-  @Override
-  public boolean removeAll(Collection<?> c) {
-    ensureIsMutable();
-    return super.removeAll(c);
-  }
-  
-  @Override
-  public boolean retainAll(Collection<?> c) {
-    ensureIsMutable();
-    return super.retainAll(c);
-  }
-  
-  @Override
-  public E set(int index, E element) {
-    ensureIsMutable();
-    return super.set(index, element);
-  }
-  
-  /**
-   * Throws an {@link UnsupportedOperationException} if the list is immutable. Subclasses are
-   * responsible for invoking this method on mutate operations.
-   */
-  protected void ensureIsMutable() {
-    if (!isMutable) {
-      throw new UnsupportedOperationException();
-    }
-  }
-}


[39/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Api.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Api.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Api.java
deleted file mode 100644
index 3265158..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Api.java
+++ /dev/null
@@ -1,2473 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Api is a light-weight descriptor for a protocol buffer service.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Api}
- */
-public  final class Api extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Api)
-    ApiOrBuilder {
-  // Use Api.newBuilder() to construct.
-  private Api(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Api() {
-    name_ = "";
-    methods_ = java.util.Collections.emptyList();
-    options_ = java.util.Collections.emptyList();
-    version_ = "";
-    mixins_ = java.util.Collections.emptyList();
-    syntax_ = 0;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Api(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            name_ = s;
-            break;
-          }
-          case 18: {
-            if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-              methods_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method>();
-              mutable_bitField0_ |= 0x00000002;
-            }
-            methods_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.parser(), extensionRegistry));
-            break;
-          }
-          case 26: {
-            if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-              options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>();
-              mutable_bitField0_ |= 0x00000004;
-            }
-            options_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.parser(), extensionRegistry));
-            break;
-          }
-          case 34: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            version_ = s;
-            break;
-          }
-          case 42: {
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder subBuilder = null;
-            if (sourceContext_ != null) {
-              subBuilder = sourceContext_.toBuilder();
-            }
-            sourceContext_ = input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.parser(), extensionRegistry);
-            if (subBuilder != null) {
-              subBuilder.mergeFrom(sourceContext_);
-              sourceContext_ = subBuilder.buildPartial();
-            }
-
-            break;
-          }
-          case 50: {
-            if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
-              mixins_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin>();
-              mutable_bitField0_ |= 0x00000020;
-            }
-            mixins_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.parser(), extensionRegistry));
-            break;
-          }
-          case 56: {
-            int rawValue = input.readEnum();
-
-            syntax_ = rawValue;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-        methods_ = java.util.Collections.unmodifiableList(methods_);
-      }
-      if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-        options_ = java.util.Collections.unmodifiableList(options_);
-      }
-      if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
-        mixins_ = java.util.Collections.unmodifiableList(mixins_);
-      }
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Api_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Api_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Api.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Api.Builder.class);
-  }
-
-  private int bitField0_;
-  public static final int NAME_FIELD_NUMBER = 1;
-  private volatile java.lang.Object name_;
-  /**
-   * <pre>
-   * The fully qualified name of this api, including package name
-   * followed by the api's simple name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public java.lang.String getName() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      name_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The fully qualified name of this api, including package name
-   * followed by the api's simple name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      name_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int METHODS_FIELD_NUMBER = 2;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method> methods_;
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method> getMethodsList() {
-    return methods_;
-  }
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder> 
-      getMethodsOrBuilderList() {
-    return methods_;
-  }
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  public int getMethodsCount() {
-    return methods_.size();
-  }
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method getMethods(int index) {
-    return methods_.get(index);
-  }
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder getMethodsOrBuilder(
-      int index) {
-    return methods_.get(index);
-  }
-
-  public static final int OPTIONS_FIELD_NUMBER = 3;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_;
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public int getOptionsCount() {
-    return options_.size();
-  }
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-    return options_.get(index);
-  }
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index) {
-    return options_.get(index);
-  }
-
-  public static final int VERSION_FIELD_NUMBER = 4;
-  private volatile java.lang.Object version_;
-  /**
-   * <pre>
-   * A version string for this api. If specified, must have the form
-   * `major-version.minor-version`, as in `1.10`. If the minor version
-   * is omitted, it defaults to zero. If the entire version field is
-   * empty, the major version is derived from the package name, as
-   * outlined below. If the field is not empty, the version in the
-   * package name will be verified to be consistent with what is
-   * provided here.
-   * The versioning schema uses [semantic
-   * versioning](http://semver.org) where the major version number
-   * indicates a breaking change and the minor version an additive,
-   * non-breaking change. Both version numbers are signals to users
-   * what to expect from different versions, and should be carefully
-   * chosen based on the product plan.
-   * The major version is also reflected in the package name of the
-   * API, which must end in `v&lt;major-version&gt;`, as in
-   * `google.feature.v1`. For major versions 0 and 1, the suffix can
-   * be omitted. Zero major versions must only be used for
-   * experimental, none-GA apis.
-   * </pre>
-   *
-   * <code>string version = 4;</code>
-   */
-  public java.lang.String getVersion() {
-    java.lang.Object ref = version_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      version_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * A version string for this api. If specified, must have the form
-   * `major-version.minor-version`, as in `1.10`. If the minor version
-   * is omitted, it defaults to zero. If the entire version field is
-   * empty, the major version is derived from the package name, as
-   * outlined below. If the field is not empty, the version in the
-   * package name will be verified to be consistent with what is
-   * provided here.
-   * The versioning schema uses [semantic
-   * versioning](http://semver.org) where the major version number
-   * indicates a breaking change and the minor version an additive,
-   * non-breaking change. Both version numbers are signals to users
-   * what to expect from different versions, and should be carefully
-   * chosen based on the product plan.
-   * The major version is also reflected in the package name of the
-   * API, which must end in `v&lt;major-version&gt;`, as in
-   * `google.feature.v1`. For major versions 0 and 1, the suffix can
-   * be omitted. Zero major versions must only be used for
-   * experimental, none-GA apis.
-   * </pre>
-   *
-   * <code>string version = 4;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getVersionBytes() {
-    java.lang.Object ref = version_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      version_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int SOURCE_CONTEXT_FIELD_NUMBER = 5;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext sourceContext_;
-  /**
-   * <pre>
-   * Source context for the protocol buffer service represented by this
-   * message.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  public boolean hasSourceContext() {
-    return sourceContext_ != null;
-  }
-  /**
-   * <pre>
-   * Source context for the protocol buffer service represented by this
-   * message.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext() {
-    return sourceContext_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-  }
-  /**
-   * <pre>
-   * Source context for the protocol buffer service represented by this
-   * message.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder() {
-    return getSourceContext();
-  }
-
-  public static final int MIXINS_FIELD_NUMBER = 6;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin> mixins_;
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin> getMixinsList() {
-    return mixins_;
-  }
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder> 
-      getMixinsOrBuilderList() {
-    return mixins_;
-  }
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  public int getMixinsCount() {
-    return mixins_.size();
-  }
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin getMixins(int index) {
-    return mixins_.get(index);
-  }
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder getMixinsOrBuilder(
-      int index) {
-    return mixins_.get(index);
-  }
-
-  public static final int SYNTAX_FIELD_NUMBER = 7;
-  private int syntax_;
-  /**
-   * <pre>
-   * The source syntax of the service.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 7;</code>
-   */
-  public int getSyntaxValue() {
-    return syntax_;
-  }
-  /**
-   * <pre>
-   * The source syntax of the service.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 7;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax() {
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.valueOf(syntax_);
-    return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.UNRECOGNIZED : result;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
-    }
-    for (int i = 0; i < methods_.size(); i++) {
-      output.writeMessage(2, methods_.get(i));
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      output.writeMessage(3, options_.get(i));
-    }
-    if (!getVersionBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, version_);
-    }
-    if (sourceContext_ != null) {
-      output.writeMessage(5, getSourceContext());
-    }
-    for (int i = 0; i < mixins_.size(); i++) {
-      output.writeMessage(6, mixins_.get(i));
-    }
-    if (syntax_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.SYNTAX_PROTO2.getNumber()) {
-      output.writeEnum(7, syntax_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
-    }
-    for (int i = 0; i < methods_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(2, methods_.get(i));
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(3, options_.get(i));
-    }
-    if (!getVersionBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, version_);
-    }
-    if (sourceContext_ != null) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(5, getSourceContext());
-    }
-    for (int i = 0; i < mixins_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(6, mixins_.get(i));
-    }
-    if (syntax_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.SYNTAX_PROTO2.getNumber()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeEnumSize(7, syntax_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Api)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Api other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Api) obj;
-
-    boolean result = true;
-    result = result && getName()
-        .equals(other.getName());
-    result = result && getMethodsList()
-        .equals(other.getMethodsList());
-    result = result && getOptionsList()
-        .equals(other.getOptionsList());
-    result = result && getVersion()
-        .equals(other.getVersion());
-    result = result && (hasSourceContext() == other.hasSourceContext());
-    if (hasSourceContext()) {
-      result = result && getSourceContext()
-          .equals(other.getSourceContext());
-    }
-    result = result && getMixinsList()
-        .equals(other.getMixinsList());
-    result = result && syntax_ == other.syntax_;
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getName().hashCode();
-    if (getMethodsCount() > 0) {
-      hash = (37 * hash) + METHODS_FIELD_NUMBER;
-      hash = (53 * hash) + getMethodsList().hashCode();
-    }
-    if (getOptionsCount() > 0) {
-      hash = (37 * hash) + OPTIONS_FIELD_NUMBER;
-      hash = (53 * hash) + getOptionsList().hashCode();
-    }
-    hash = (37 * hash) + VERSION_FIELD_NUMBER;
-    hash = (53 * hash) + getVersion().hashCode();
-    if (hasSourceContext()) {
-      hash = (37 * hash) + SOURCE_CONTEXT_FIELD_NUMBER;
-      hash = (53 * hash) + getSourceContext().hashCode();
-    }
-    if (getMixinsCount() > 0) {
-      hash = (37 * hash) + MIXINS_FIELD_NUMBER;
-      hash = (53 * hash) + getMixinsList().hashCode();
-    }
-    hash = (37 * hash) + SYNTAX_FIELD_NUMBER;
-    hash = (53 * hash) + syntax_;
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Api prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Api is a light-weight descriptor for a protocol buffer service.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Api}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Api)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Api_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Api_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Api.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Api.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Api.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-        getMethodsFieldBuilder();
-        getOptionsFieldBuilder();
-        getMixinsFieldBuilder();
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      name_ = "";
-
-      if (methodsBuilder_ == null) {
-        methods_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000002);
-      } else {
-        methodsBuilder_.clear();
-      }
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000004);
-      } else {
-        optionsBuilder_.clear();
-      }
-      version_ = "";
-
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = null;
-      } else {
-        sourceContext_ = null;
-        sourceContextBuilder_ = null;
-      }
-      if (mixinsBuilder_ == null) {
-        mixins_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000020);
-      } else {
-        mixinsBuilder_.clear();
-      }
-      syntax_ = 0;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Api_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Api getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Api.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Api build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Api result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Api buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Api result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Api(this);
-      int from_bitField0_ = bitField0_;
-      int to_bitField0_ = 0;
-      result.name_ = name_;
-      if (methodsBuilder_ == null) {
-        if (((bitField0_ & 0x00000002) == 0x00000002)) {
-          methods_ = java.util.Collections.unmodifiableList(methods_);
-          bitField0_ = (bitField0_ & ~0x00000002);
-        }
-        result.methods_ = methods_;
-      } else {
-        result.methods_ = methodsBuilder_.build();
-      }
-      if (optionsBuilder_ == null) {
-        if (((bitField0_ & 0x00000004) == 0x00000004)) {
-          options_ = java.util.Collections.unmodifiableList(options_);
-          bitField0_ = (bitField0_ & ~0x00000004);
-        }
-        result.options_ = options_;
-      } else {
-        result.options_ = optionsBuilder_.build();
-      }
-      result.version_ = version_;
-      if (sourceContextBuilder_ == null) {
-        result.sourceContext_ = sourceContext_;
-      } else {
-        result.sourceContext_ = sourceContextBuilder_.build();
-      }
-      if (mixinsBuilder_ == null) {
-        if (((bitField0_ & 0x00000020) == 0x00000020)) {
-          mixins_ = java.util.Collections.unmodifiableList(mixins_);
-          bitField0_ = (bitField0_ & ~0x00000020);
-        }
-        result.mixins_ = mixins_;
-      } else {
-        result.mixins_ = mixinsBuilder_.build();
-      }
-      result.syntax_ = syntax_;
-      result.bitField0_ = to_bitField0_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Api) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Api)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Api other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Api.getDefaultInstance()) return this;
-      if (!other.getName().isEmpty()) {
-        name_ = other.name_;
-        onChanged();
-      }
-      if (methodsBuilder_ == null) {
-        if (!other.methods_.isEmpty()) {
-          if (methods_.isEmpty()) {
-            methods_ = other.methods_;
-            bitField0_ = (bitField0_ & ~0x00000002);
-          } else {
-            ensureMethodsIsMutable();
-            methods_.addAll(other.methods_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.methods_.isEmpty()) {
-          if (methodsBuilder_.isEmpty()) {
-            methodsBuilder_.dispose();
-            methodsBuilder_ = null;
-            methods_ = other.methods_;
-            bitField0_ = (bitField0_ & ~0x00000002);
-            methodsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getMethodsFieldBuilder() : null;
-          } else {
-            methodsBuilder_.addAllMessages(other.methods_);
-          }
-        }
-      }
-      if (optionsBuilder_ == null) {
-        if (!other.options_.isEmpty()) {
-          if (options_.isEmpty()) {
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000004);
-          } else {
-            ensureOptionsIsMutable();
-            options_.addAll(other.options_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.options_.isEmpty()) {
-          if (optionsBuilder_.isEmpty()) {
-            optionsBuilder_.dispose();
-            optionsBuilder_ = null;
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000004);
-            optionsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getOptionsFieldBuilder() : null;
-          } else {
-            optionsBuilder_.addAllMessages(other.options_);
-          }
-        }
-      }
-      if (!other.getVersion().isEmpty()) {
-        version_ = other.version_;
-        onChanged();
-      }
-      if (other.hasSourceContext()) {
-        mergeSourceContext(other.getSourceContext());
-      }
-      if (mixinsBuilder_ == null) {
-        if (!other.mixins_.isEmpty()) {
-          if (mixins_.isEmpty()) {
-            mixins_ = other.mixins_;
-            bitField0_ = (bitField0_ & ~0x00000020);
-          } else {
-            ensureMixinsIsMutable();
-            mixins_.addAll(other.mixins_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.mixins_.isEmpty()) {
-          if (mixinsBuilder_.isEmpty()) {
-            mixinsBuilder_.dispose();
-            mixinsBuilder_ = null;
-            mixins_ = other.mixins_;
-            bitField0_ = (bitField0_ & ~0x00000020);
-            mixinsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getMixinsFieldBuilder() : null;
-          } else {
-            mixinsBuilder_.addAllMessages(other.mixins_);
-          }
-        }
-      }
-      if (other.syntax_ != 0) {
-        setSyntaxValue(other.getSyntaxValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Api parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Api) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private java.lang.Object name_ = "";
-    /**
-     * <pre>
-     * The fully qualified name of this api, including package name
-     * followed by the api's simple name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        name_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The fully qualified name of this api, including package name
-     * followed by the api's simple name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The fully qualified name of this api, including package name
-     * followed by the api's simple name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      name_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The fully qualified name of this api, including package name
-     * followed by the api's simple name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder clearName() {
-      
-      name_ = getDefaultInstance().getName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The fully qualified name of this api, including package name
-     * followed by the api's simple name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      name_ = value;
-      onChanged();
-      return this;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method> methods_ =
-      java.util.Collections.emptyList();
-    private void ensureMethodsIsMutable() {
-      if (!((bitField0_ & 0x00000002) == 0x00000002)) {
-        methods_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method>(methods_);
-        bitField0_ |= 0x00000002;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Method, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder> methodsBuilder_;
-
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method> getMethodsList() {
-      if (methodsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(methods_);
-      } else {
-        return methodsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public int getMethodsCount() {
-      if (methodsBuilder_ == null) {
-        return methods_.size();
-      } else {
-        return methodsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method getMethods(int index) {
-      if (methodsBuilder_ == null) {
-        return methods_.get(index);
-      } else {
-        return methodsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder setMethods(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method value) {
-      if (methodsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureMethodsIsMutable();
-        methods_.set(index, value);
-        onChanged();
-      } else {
-        methodsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder setMethods(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder builderForValue) {
-      if (methodsBuilder_ == null) {
-        ensureMethodsIsMutable();
-        methods_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        methodsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder addMethods(org.apache.hadoop.hbase.shaded.com.google.protobuf.Method value) {
-      if (methodsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureMethodsIsMutable();
-        methods_.add(value);
-        onChanged();
-      } else {
-        methodsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder addMethods(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method value) {
-      if (methodsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureMethodsIsMutable();
-        methods_.add(index, value);
-        onChanged();
-      } else {
-        methodsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder addMethods(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder builderForValue) {
-      if (methodsBuilder_ == null) {
-        ensureMethodsIsMutable();
-        methods_.add(builderForValue.build());
-        onChanged();
-      } else {
-        methodsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder addMethods(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder builderForValue) {
-      if (methodsBuilder_ == null) {
-        ensureMethodsIsMutable();
-        methods_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        methodsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder addAllMethods(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Method> values) {
-      if (methodsBuilder_ == null) {
-        ensureMethodsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, methods_);
-        onChanged();
-      } else {
-        methodsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder clearMethods() {
-      if (methodsBuilder_ == null) {
-        methods_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000002);
-        onChanged();
-      } else {
-        methodsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public Builder removeMethods(int index) {
-      if (methodsBuilder_ == null) {
-        ensureMethodsIsMutable();
-        methods_.remove(index);
-        onChanged();
-      } else {
-        methodsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder getMethodsBuilder(
-        int index) {
-      return getMethodsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder getMethodsOrBuilder(
-        int index) {
-      if (methodsBuilder_ == null) {
-        return methods_.get(index);  } else {
-        return methodsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder> 
-         getMethodsOrBuilderList() {
-      if (methodsBuilder_ != null) {
-        return methodsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(methods_);
-      }
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder addMethodsBuilder() {
-      return getMethodsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder addMethodsBuilder(
-        int index) {
-      return getMethodsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * The methods of this api, in unspecified order.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Method methods = 2;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder> 
-         getMethodsBuilderList() {
-      return getMethodsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Method, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder> 
-        getMethodsFieldBuilder() {
-      if (methodsBuilder_ == null) {
-        methodsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Method, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder>(
-                methods_,
-                ((bitField0_ & 0x00000002) == 0x00000002),
-                getParentForChildren(),
-                isClean());
-        methods_ = null;
-      }
-      return methodsBuilder_;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_ =
-      java.util.Collections.emptyList();
-    private void ensureOptionsIsMutable() {
-      if (!((bitField0_ & 0x00000004) == 0x00000004)) {
-        options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>(options_);
-        bitField0_ |= 0x00000004;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> optionsBuilder_;
-
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-      if (optionsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(options_);
-      } else {
-        return optionsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public int getOptionsCount() {
-      if (optionsBuilder_ == null) {
-        return options_.size();
-      } else {
-        return optionsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);
-      } else {
-        return optionsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.set(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addAllOptions(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> values) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, options_);
-        onChanged();
-      } else {
-        optionsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder clearOptions() {
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000004);
-        onChanged();
-      } else {
-        optionsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder removeOptions(int index) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.remove(index);
-        onChanged();
-      } else {
-        optionsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder getOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-        int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);  } else {
-        return optionsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-         getOptionsOrBuilderList() {
-      if (optionsBuilder_ != null) {
-        return optionsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(options_);
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder() {
-      return getOptionsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the API.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder> 
-         getOptionsBuilderList() {
-      return getOptionsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-        getOptionsFieldBuilder() {
-      if (optionsBuilder_ == null) {
-        optionsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder>(
-                options_,
-                ((bitField0_ & 0x00000004) == 0x00000004),
-                getParentForChildren(),
-                isClean());
-        options_ = null;
-      }
-      return optionsBuilder_;
-    }
-
-    private java.lang.Object version_ = "";
-    /**
-     * <pre>
-     * A version string for this api. If specified, must have the form
-     * `major-version.minor-version`, as in `1.10`. If the minor version
-     * is omitted, it defaults to zero. If the entire version field is
-     * empty, the major version is derived from the package name, as
-     * outlined below. If the field is not empty, the version in the
-     * package name will be verified to be consistent with what is
-     * provided here.
-     * The versioning schema uses [semantic
-     * versioning](http://semver.org) where the major version number
-     * indicates a breaking change and the minor version an additive,
-     * non-breaking change. Both version numbers are signals to users
-     * what to expect from different versions, and should be carefully
-     * chosen based on the product plan.
-     * The major version is also reflected in the package name of the
-     * API, which must end in `v&lt;major-version&gt;`, as in
-     * `google.feature.v1`. For major versions 0 and 1, the suffix can
-     * be omitted. Zero major versions must only be used for
-     * experimental, none-GA apis.
-     * </pre>
-     *
-     * <code>string version = 4;</code>
-     */
-    public java.lang.String getVersion() {
-      java.lang.Object ref = version_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        version_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * A version string for this api. If specified, must have the form
-     * `major-version.minor-version`, as in `1.10`. If the minor version
-     * is omitted, it defaults to zero. If the entire version field is
-     * empty, the major version is derived from the package name, as
-     * outlined below. If the field is not empty, the version in the
-     * package name will be verified to be consistent with what is
-     * provided here.
-     * The versioning schema uses [semantic
-     * versioning](http://semver.org) where the major version number
-     * indicates a breaking change and the minor version an additive,
-     * non-breaking change. Both version numbers are signals to users
-     * what to expect from different versions, and should be carefully
-     * chosen based on the product plan.
-     * The major version is also reflected in the package name of the
-     * API, which must end in `v&lt;major-version&gt;`, as in
-     * `google.feature.v1`. For major versions 0 and 1, the suffix can
-     * be omitted. Zero major versions must only be used for
-     * experimental, none-GA apis.
-     * </pre>
-     *
-     * <code>string version = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getVersionBytes() {
-      java.lang.Object ref = version_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        version_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * A version string for this api. If specified, must have the form
-     * `major-version.minor-version`, as in `1.10`. If the minor version
-     * is omitted, it defaults to zero. If the entire version field is
-     * empty, the major version is derived from the package name, as
-     * outlined below. If the field is not empty, the version in the
-     * package name will be verified to be consistent with what is
-     * provided here.
-     * The versioning schema uses [semantic
-     * versioning](http://semver.org) where the major version number
-     * indicates a breaking change and the minor version an additive,
-     * non-breaking change. Both version numbers are signals to users
-     * what to expect from different versions, and should be carefully
-     * chosen based on the product plan.
-     * The major version is also reflected in the package name of the
-     * API, which must end in `v&lt;major-version&gt;`, as in
-     * `google.feature.v1`. For major versions 0 and 1, the suffix can
-     * be omitted. Zero major versions must only be used for
-     * experimental, none-GA apis.
-     * </pre>
-     *
-     * <code>string version = 4;</code>
-     */
-    public Builder setVersion(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      version_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * A version string for this api. If specified, must have the form
-     * `major-version.minor-version`, as in `1.10`. If the minor version
-     * is omitted, it defaults to zero. If the entire version field is
-     * empty, the major version is derived from the package name, as
-     * outlined below. If the field is not empty, the version in the
-     * package name will be verified to be consistent with what is
-     * provided here.
-     * The versioning schema uses [semantic
-     * versioning](http://semver.org) where the major version number
-     * indicates a breaking change and the minor version an additive,
-     * non-breaking change. Both version numbers are signals to users
-     * what to expect from different versions, and should be carefully
-     * chosen based on the product plan.
-     * The major version is also reflected in the package name of the
-     * API, which must end in `v&lt;major-version&gt;`, as in
-     * `google.feature.v1`. For major versions 0 and 1, the suffix can
-     * be omitted. Zero major versions must only be used for
-     * experimental, none-GA apis.
-     * </pre>
-     *
-     * <code>string version = 4;</code>
-     */
-    public Builder clearVersion() {
-      
-      version_ = getDefaultInstance().getVersion();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * A version string for this api. If specified, must have the form
-     * `major-version.minor-version`, as in `1.10`. If the minor version
-     * is omitted, it defaults to zero. If the entire version field is
-     * empty, the major version is derived from the package name, as
-     * outlined below. If the field is not empty, the version in the
-     * package name will be verified to be consistent with what is
-     * provided here.
-     * The versioning schema uses [semantic
-     * versioning](http://semver.org) where the major version number
-     * indicates a breaking change and the minor version an additive,
-     * non-breaking change. Both version numbers are signals to users
-     * what to expect from different versions, and should be carefully
-     * chosen based on the product plan.
-     * The major version is also reflected in the package name of the
-     * API, which must end in `v&lt;major-version&gt;`, as in
-     * `google.feature.v1`. For major versions 0 and 1, the suffix can
-     * be omitted. Zero major versions must only be used for
-     * experimental, none-GA apis.
-     * </pre>
-     *
-     * <code>string version = 4;</code>
-     */
-    public Builder setVersionBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      version_ = value;
-      onChanged();
-      return this;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext sourceContext_ = null;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder> sourceContextBuilder_;
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public boolean hasSourceContext() {
-      return sourceContextBuilder_ != null || sourceContext_ != null;
-    }
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext() {
-      if (sourceContextBuilder_ == null) {
-        return sourceContext_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-      } else {
-        return sourceContextBuilder_.getMessage();
-      }
-    }
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public Builder setSourceContext(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext value) {
-      if (sourceContextBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        sourceContext_ = value;
-        onChanged();
-      } else {
-        sourceContextBuilder_.setMessage(value);
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public Builder setSourceContext(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder builderForValue) {
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = builderForValue.build();
-        onChanged();
-      } else {
-        sourceContextBuilder_.setMessage(builderForValue.build());
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public Builder mergeSourceContext(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext value) {
-      if (sourceContextBuilder_ == null) {
-        if (sourceContext_ != null) {
-          sourceContext_ =
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.newBuilder(sourceContext_).mergeFrom(value).buildPartial();
-        } else {
-          sourceContext_ = value;
-        }
-        onChanged();
-      } else {
-        sourceContextBuilder_.mergeFrom(value);
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public Builder clearSourceContext() {
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = null;
-        onChanged();
-      } else {
-        sourceContext_ = null;
-        sourceContextBuilder_ = null;
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder getSourceContextBuilder() {
-      
-      onChanged();
-      return getSourceContextFieldBuilder().getBuilder();
-    }
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder() {
-      if (sourceContextBuilder_ != null) {
-        return sourceContextBuilder_.getMessageOrBuilder();
-      } else {
-        return sourceContext_ == null ?
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-      }
-    }
-    /**
-     * <pre>
-     * Source context for the protocol buffer service represented by this
-     * message.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 5;</code>
-     */
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder> 
-        getSourceContextFieldBuilder() {
-      if (sourceContextBuilder_ == null) {
-        sourceContextBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder>(
-                getSourceContext(),
-                getParentForChildren(),
-                isClean());
-        sourceContext_ = null;
-      }
-      return sourceContextBuilder_;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin> mixins_ =
-      java.util.Collections.emptyList();
-    private void ensureMixinsIsMutable() {
-      if (!((bitField0_ & 0x00000020) == 0x00000020)) {
-        mixins_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin>(mixins_);
-        bitField0_ |= 0x00000020;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder> mixinsBuilder_;
-
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin> getMixinsList() {
-      if (mixinsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(mixins_);
-      } else {
-        return mixinsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public int getMixinsCount() {
-      if (mixinsBuilder_ == null) {
-        return mixins_.size();
-      } else {
-        return mixinsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin getMixins(int index) {
-      if (mixinsBuilder_ == null) {
-        return mixins_.get(index);
-      } else {
-        return mixinsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder setMixins(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin value) {
-      if (mixinsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureMixinsIsMutable();
-        mixins_.set(index, value);
-        onChanged();
-      } else {
-        mixinsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder setMixins(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder builderForValue) {
-      if (mixinsBuilder_ == null) {
-        ensureMixinsIsMutable();
-        mixins_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        mixinsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder addMixins(org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin value) {
-      if (mixinsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureMixinsIsMutable();
-        mixins_.add(value);
-        onChanged();
-      } else {
-        mixinsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder addMixins(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin value) {
-      if (mixinsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureMixinsIsMutable();
-        mixins_.add(index, value);
-        onChanged();
-      } else {
-        mixinsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder addMixins(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder builderForValue) {
-      if (mixinsBuilder_ == null) {
-        ensureMixinsIsMutable();
-        mixins_.add(builderForValue.build());
-        onChanged();
-      } else {
-        mixinsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder addMixins(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder builderForValue) {
-      if (mixinsBuilder_ == null) {
-        ensureMixinsIsMutable();
-        mixins_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        mixinsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder addAllMixins(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin> values) {
-      if (mixinsBuilder_ == null) {
-        ensureMixinsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, mixins_);
-        onChanged();
-      } else {
-        mixinsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder clearMixins() {
-      if (mixinsBuilder_ == null) {
-        mixins_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000020);
-        onChanged();
-      } else {
-        mixinsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public Builder removeMixins(int index) {
-      if (mixinsBuilder_ == null) {
-        ensureMixinsIsMutable();
-        mixins_.remove(index);
-        onChanged();
-      } else {
-        mixinsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder getMixinsBuilder(
-        int index) {
-      return getMixinsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder getMixinsOrBuilder(
-        int index) {
-      if (mixinsBuilder_ == null) {
-        return mixins_.get(index);  } else {
-        return mixinsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder> 
-         getMixinsOrBuilderList() {
-      if (mixinsBuilder_ != null) {
-        return mixinsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(mixins_);
-      }
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder addMixinsBuilder() {
-      return getMixinsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder addMixinsBuilder(
-        int index) {
-      return getMixinsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Included APIs. See [Mixin][].
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder> 
-         getMixinsBuilderList() {
-      return getMixinsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder> 
-        getMixinsFieldBuilder() {
-      if (mixinsBuilder_ == null) {
-        mixinsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder>(
-                mixins_,
-                ((bitField0_ & 0x00000020) == 0x00000020),
-                getParentForChildren(),
-                isClean());
-        mixins_ = null;
-      }
-      return mixinsBuilder_;
-    }
-
-    private int syntax_ = 0;
-    /**
-     * <pre>
-     * The source syntax of the service.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public int getSyntaxValue() {
-      return syntax_;
-    }
-    /**
-     * <pre>
-     * The source syntax of the service.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public Builder setSyntaxValue(int value) {
-      syntax_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The source syntax of the service.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.valueOf(syntax_);
-      return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.UNRECOGNIZED : result;
-    }
-    /**
-     * <pre>
-     * The source syntax of the service.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public Builder setSyntax(org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax value) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-      
-      syntax_ = value.getNumber();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The source syntax of the service.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public Builder clearSyntax() {
-      
-      syntax_ = 0;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Api)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Api)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Api DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Api();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Api getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Api>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Api>() {
-    public Api parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Api(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Api> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Api> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Api getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-


[36/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/CodedInputStream.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/CodedInputStream.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/CodedInputStream.java
deleted file mode 100644
index 23cc1a4..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/CodedInputStream.java
+++ /dev/null
@@ -1,3549 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EMPTY_BYTE_ARRAY;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EMPTY_BYTE_BUFFER;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.UTF_8;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.checkNotNull;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.FIXED_32_SIZE;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.FIXED_64_SIZE;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.MAX_VARINT_SIZE;
-
-import java.io.ByteArrayOutputStream;
-import java.io.IOException;
-import java.io.InputStream;
-import java.nio.ByteBuffer;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.List;
-
-/**
- * Reads and decodes protocol message fields.
- *
- * <p>This class contains two kinds of methods: methods that read specific protocol message
- * constructs and field types (e.g. {@link #readTag()} and {@link #readInt32()}) and methods that
- * read low-level values (e.g. {@link #readRawVarint32()} and {@link #readRawBytes}). If you are
- * reading encoded protocol messages, you should use the former methods, but if you are reading some
- * other format of your own design, use the latter.
- *
- * @author kenton@google.com Kenton Varda
- */
-public abstract class CodedInputStream {
-  private static final int DEFAULT_BUFFER_SIZE = 4096;
-  private static final int DEFAULT_RECURSION_LIMIT = 100;
-  // Integer.MAX_VALUE == 0x7FFFFFF == INT_MAX from limits.h
-  private static final int DEFAULT_SIZE_LIMIT = Integer.MAX_VALUE;
-
-  /** Visible for subclasses. See setRecursionLimit() */
-  int recursionDepth;
-
-  int recursionLimit = DEFAULT_RECURSION_LIMIT;
-
-  /** Visible for subclasses. See setSizeLimit() */
-  int sizeLimit = DEFAULT_SIZE_LIMIT;
-
-  /** Create a new CodedInputStream wrapping the given InputStream. */
-  public static CodedInputStream newInstance(final InputStream input) {
-    return newInstance(input, DEFAULT_BUFFER_SIZE);
-  }
-
-  /** Create a new CodedInputStream wrapping the given InputStream. */
-  static CodedInputStream newInstance(final InputStream input, int bufferSize) {
-    if (input == null) {
-      // TODO(nathanmittler): Ideally we should throw here. This is done for backward compatibility.
-      return newInstance(EMPTY_BYTE_ARRAY);
-    }
-    return new StreamDecoder(input, bufferSize);
-  }
-
-  /** Create a new CodedInputStream wrapping the given byte array. */
-  public static CodedInputStream newInstance(final byte[] buf) {
-    return newInstance(buf, 0, buf.length);
-  }
-
-  /** Create a new CodedInputStream wrapping the given byte array slice. */
-  public static CodedInputStream newInstance(final byte[] buf, final int off, final int len) {
-    return newInstance(buf, off, len, false /* bufferIsImmutable */);
-  }
-
-  /** Create a new CodedInputStream wrapping the given byte array slice. */
-  static CodedInputStream newInstance(
-      final byte[] buf, final int off, final int len, final boolean bufferIsImmutable) {
-    ArrayDecoder result = new ArrayDecoder(buf, off, len, bufferIsImmutable);
-    try {
-      // Some uses of CodedInputStream can be more efficient if they know
-      // exactly how many bytes are available.  By pushing the end point of the
-      // buffer as a limit, we allow them to get this information via
-      // getBytesUntilLimit().  Pushing a limit that we know is at the end of
-      // the stream can never hurt, since we can never past that point anyway.
-      result.pushLimit(len);
-    } catch (InvalidProtocolBufferException ex) {
-      // The only reason pushLimit() might throw an exception here is if len
-      // is negative. Normally pushLimit()'s parameter comes directly off the
-      // wire, so it's important to catch exceptions in case of corrupt or
-      // malicious data. However, in this case, we expect that len is not a
-      // user-supplied value, so we can assume that it being negative indicates
-      // a programming error. Therefore, throwing an unchecked exception is
-      // appropriate.
-      throw new IllegalArgumentException(ex);
-    }
-    return result;
-  }
-
-  /**
-   * Create a new CodedInputStream wrapping the given ByteBuffer. The data starting from the
-   * ByteBuffer's current position to its limit will be read. The returned CodedInputStream may or
-   * may not share the underlying data in the ByteBuffer, therefore the ByteBuffer cannot be changed
-   * while the CodedInputStream is in use. Note that the ByteBuffer's position won't be changed by
-   * this function. Concurrent calls with the same ByteBuffer object are safe if no other thread is
-   * trying to alter the ByteBuffer's status.
-   */
-  public static CodedInputStream newInstance(ByteBuffer buf) {
-    return newInstance(buf, false /* bufferIsImmutable */);
-  }
-
-  /** Create a new CodedInputStream wrapping the given buffer. */
-  static CodedInputStream newInstance(ByteBuffer buf, boolean bufferIsImmutable) {
-    if (buf.hasArray()) {
-      return newInstance(
-          buf.array(), buf.arrayOffset() + buf.position(), buf.remaining(), bufferIsImmutable);
-    }
-
-    if (buf.isDirect() && UnsafeDirectNioDecoder.isSupported()) {
-      return new UnsafeDirectNioDecoder(buf, bufferIsImmutable);
-    }
-
-    // The buffer is non-direct and does not expose the underlying array. Using the ByteBuffer API
-    // to access individual bytes is very slow, so just copy the buffer to an array.
-    // TODO(nathanmittler): Re-evaluate with Java 9
-    byte[] buffer = new byte[buf.remaining()];
-    buf.duplicate().get(buffer);
-    return newInstance(buffer, 0, buffer.length, true);
-  }
-
-  /** Create a new CodedInputStream wrapping the given {@link ByteInput}. */
-  static CodedInputStream newInstance(ByteInput buf, int off, int len, boolean bufferIsImmutable) {
-    return new ByteInputDecoder(buf, off, len, bufferIsImmutable);
-  }
-
-  /** Disable construction/inheritance outside of this class. */
-  private CodedInputStream() {}
-
-  // -----------------------------------------------------------------
-
-  /**
-   * Attempt to read a field tag, returning zero if we have reached EOF. Protocol message parsers
-   * use this to read tags, since a protocol message may legally end wherever a tag occurs, and zero
-   * is not a valid tag number.
-   */
-  public abstract int readTag() throws IOException;
-
-  /**
-   * Verifies that the last call to readTag() returned the given tag value. This is used to verify
-   * that a nested group ended with the correct end tag.
-   *
-   * @throws InvalidProtocolBufferException {@code value} does not match the last tag.
-   */
-  public abstract void checkLastTagWas(final int value) throws InvalidProtocolBufferException;
-
-  public abstract int getLastTag();
-
-  /**
-   * Reads and discards a single field, given its tag value.
-   *
-   * @return {@code false} if the tag is an endgroup tag, in which case nothing is skipped.
-   *     Otherwise, returns {@code true}.
-   */
-  public abstract boolean skipField(final int tag) throws IOException;
-
-  /**
-   * Reads a single field and writes it to output in wire format, given its tag value.
-   *
-   * @return {@code false} if the tag is an endgroup tag, in which case nothing is skipped.
-   *     Otherwise, returns {@code true}.
-   * @deprecated use {@code UnknownFieldSet} or {@code UnknownFieldSetLite} to skip to an output
-   *     stream.
-   */
-  @Deprecated
-  public abstract boolean skipField(final int tag, final CodedOutputStream output)
-      throws IOException;
-
-  /**
-   * Reads and discards an entire message. This will read either until EOF or until an endgroup tag,
-   * whichever comes first.
-   */
-  public abstract void skipMessage() throws IOException;
-
-  /**
-   * Reads an entire message and writes it to output in wire format. This will read either until EOF
-   * or until an endgroup tag, whichever comes first.
-   */
-  public abstract void skipMessage(CodedOutputStream output) throws IOException;
-
-
-  // -----------------------------------------------------------------
-
-  /** Read a {@code double} field value from the stream. */
-  public abstract double readDouble() throws IOException;
-
-  /** Read a {@code float} field value from the stream. */
-  public abstract float readFloat() throws IOException;
-
-  /** Read a {@code uint64} field value from the stream. */
-  public abstract long readUInt64() throws IOException;
-
-  /** Read an {@code int64} field value from the stream. */
-  public abstract long readInt64() throws IOException;
-
-  /** Read an {@code int32} field value from the stream. */
-  public abstract int readInt32() throws IOException;
-
-  /** Read a {@code fixed64} field value from the stream. */
-  public abstract long readFixed64() throws IOException;
-
-  /** Read a {@code fixed32} field value from the stream. */
-  public abstract int readFixed32() throws IOException;
-
-  /** Read a {@code bool} field value from the stream. */
-  public abstract boolean readBool() throws IOException;
-
-  /**
-   * Read a {@code string} field value from the stream. If the stream contains malformed UTF-8,
-   * replace the offending bytes with the standard UTF-8 replacement character.
-   */
-  public abstract String readString() throws IOException;
-
-  /**
-   * Read a {@code string} field value from the stream. If the stream contains malformed UTF-8,
-   * throw exception {@link InvalidProtocolBufferException}.
-   */
-  public abstract String readStringRequireUtf8() throws IOException;
-
-  /** Read a {@code group} field value from the stream. */
-  public abstract void readGroup(
-      final int fieldNumber,
-      final MessageLite.Builder builder,
-      final ExtensionRegistryLite extensionRegistry)
-      throws IOException;
-
-
-  /** Read a {@code group} field value from the stream. */
-  public abstract <T extends MessageLite> T readGroup(
-      final int fieldNumber, final Parser<T> parser, final ExtensionRegistryLite extensionRegistry)
-      throws IOException;
-
-  /**
-   * Reads a {@code group} field value from the stream and merges it into the given {@link
-   * UnknownFieldSet}.
-   *
-   * @deprecated UnknownFieldSet.Builder now implements MessageLite.Builder, so you can just call
-   *     {@link #readGroup}.
-   */
-  @Deprecated
-  public abstract void readUnknownGroup(final int fieldNumber, final MessageLite.Builder builder)
-      throws IOException;
-
-  /** Read an embedded message field value from the stream. */
-  public abstract void readMessage(
-      final MessageLite.Builder builder, final ExtensionRegistryLite extensionRegistry)
-      throws IOException;
-
-
-  /** Read an embedded message field value from the stream. */
-  public abstract <T extends MessageLite> T readMessage(
-      final Parser<T> parser, final ExtensionRegistryLite extensionRegistry) throws IOException;
-
-  /** Read a {@code bytes} field value from the stream. */
-  public abstract ByteString readBytes() throws IOException;
-
-  /** Read a {@code bytes} field value from the stream. */
-  public abstract byte[] readByteArray() throws IOException;
-
-  /** Read a {@code bytes} field value from the stream. */
-  public abstract ByteBuffer readByteBuffer() throws IOException;
-
-  /** Read a {@code uint32} field value from the stream. */
-  public abstract int readUInt32() throws IOException;
-
-  /**
-   * Read an enum field value from the stream. Caller is responsible for converting the numeric
-   * value to an actual enum.
-   */
-  public abstract int readEnum() throws IOException;
-
-  /** Read an {@code sfixed32} field value from the stream. */
-  public abstract int readSFixed32() throws IOException;
-
-  /** Read an {@code sfixed64} field value from the stream. */
-  public abstract long readSFixed64() throws IOException;
-
-  /** Read an {@code sint32} field value from the stream. */
-  public abstract int readSInt32() throws IOException;
-
-  /** Read an {@code sint64} field value from the stream. */
-  public abstract long readSInt64() throws IOException;
-
-  // =================================================================
-
-  /** Read a raw Varint from the stream. If larger than 32 bits, discard the upper bits. */
-  public abstract int readRawVarint32() throws IOException;
-
-  /** Read a raw Varint from the stream. */
-  public abstract long readRawVarint64() throws IOException;
-
-  /** Variant of readRawVarint64 for when uncomfortably close to the limit. */
-  /* Visible for testing */
-  abstract long readRawVarint64SlowPath() throws IOException;
-
-  /** Read a 32-bit little-endian integer from the stream. */
-  public abstract int readRawLittleEndian32() throws IOException;
-
-  /** Read a 64-bit little-endian integer from the stream. */
-  public abstract long readRawLittleEndian64() throws IOException;
-
-  // -----------------------------------------------------------------
-
-  /**
-   * Enables {@link ByteString} aliasing of the underlying buffer, trading off on buffer pinning for
-   * data copies. Only valid for buffer-backed streams.
-   */
-  public abstract void enableAliasing(boolean enabled);
-
-  /**
-   * Set the maximum message recursion depth. In order to prevent malicious messages from causing
-   * stack overflows, {@code CodedInputStream} limits how deeply messages may be nested. The default
-   * limit is 64.
-   *
-   * @return the old limit.
-   */
-  public final int setRecursionLimit(final int limit) {
-    if (limit < 0) {
-      throw new IllegalArgumentException("Recursion limit cannot be negative: " + limit);
-    }
-    final int oldLimit = recursionLimit;
-    recursionLimit = limit;
-    return oldLimit;
-  }
-
-  /**
-   * Only valid for {@link InputStream}-backed streams.
-   *
-   * <p>Set the maximum message size. In order to prevent malicious messages from exhausting memory
-   * or causing integer overflows, {@code CodedInputStream} limits how large a message may be. The
-   * default limit is 64MB. You should set this limit as small as you can without harming your app's
-   * functionality. Note that size limits only apply when reading from an {@code InputStream}, not
-   * when constructed around a raw byte array (nor with {@link ByteString#newCodedInput}).
-   *
-   * <p>If you want to read several messages from a single CodedInputStream, you could call {@link
-   * #resetSizeCounter()} after each one to avoid hitting the size limit.
-   *
-   * @return the old limit.
-   */
-  public final int setSizeLimit(final int limit) {
-    if (limit < 0) {
-      throw new IllegalArgumentException("Size limit cannot be negative: " + limit);
-    }
-    final int oldLimit = sizeLimit;
-    sizeLimit = limit;
-    return oldLimit;
-  }
-
-  /**
-   * Resets the current size counter to zero (see {@link #setSizeLimit(int)}). Only valid for {@link
-   * InputStream}-backed streams.
-   */
-  public abstract void resetSizeCounter();
-
-  /**
-   * Sets {@code currentLimit} to (current position) + {@code byteLimit}. This is called when
-   * descending into a length-delimited embedded message.
-   *
-   * <p>Note that {@code pushLimit()} does NOT affect how many bytes the {@code CodedInputStream}
-   * reads from an underlying {@code InputStream} when refreshing its buffer. If you need to prevent
-   * reading past a certain point in the underlying {@code InputStream} (e.g. because you expect it
-   * to contain more data after the end of the message which you need to handle differently) then
-   * you must place a wrapper around your {@code InputStream} which limits the amount of data that
-   * can be read from it.
-   *
-   * @return the old limit.
-   */
-  public abstract int pushLimit(int byteLimit) throws InvalidProtocolBufferException;
-
-  /**
-   * Discards the current limit, returning to the previous limit.
-   *
-   * @param oldLimit The old limit, as returned by {@code pushLimit}.
-   */
-  public abstract void popLimit(final int oldLimit);
-
-  /**
-   * Returns the number of bytes to be read before the current limit. If no limit is set, returns
-   * -1.
-   */
-  public abstract int getBytesUntilLimit();
-
-  /**
-   * Returns true if the stream has reached the end of the input. This is the case if either the end
-   * of the underlying input source has been reached or if the stream has reached a limit created
-   * using {@link #pushLimit(int)}.
-   */
-  public abstract boolean isAtEnd() throws IOException;
-
-  /**
-   * The total bytes read up to the current position. Calling {@link #resetSizeCounter()} resets
-   * this value to zero.
-   */
-  public abstract int getTotalBytesRead();
-
-  /**
-   * Read one byte from the input.
-   *
-   * @throws InvalidProtocolBufferException The end of the stream or the current limit was reached.
-   */
-  public abstract byte readRawByte() throws IOException;
-
-  /**
-   * Read a fixed size of bytes from the input.
-   *
-   * @throws InvalidProtocolBufferException The end of the stream or the current limit was reached.
-   */
-  public abstract byte[] readRawBytes(final int size) throws IOException;
-
-  /**
-   * Reads and discards {@code size} bytes.
-   *
-   * @throws InvalidProtocolBufferException The end of the stream or the current limit was reached.
-   */
-  public abstract void skipRawBytes(final int size) throws IOException;
-
-  /**
-   * Decode a ZigZag-encoded 32-bit value. ZigZag encodes signed integers into values that can be
-   * efficiently encoded with varint. (Otherwise, negative values must be sign-extended to 64 bits
-   * to be varint encoded, thus always taking 10 bytes on the wire.)
-   *
-   * @param n An unsigned 32-bit integer, stored in a signed int because Java has no explicit
-   *     unsigned support.
-   * @return A signed 32-bit integer.
-   */
-  public static int decodeZigZag32(final int n) {
-    return (n >>> 1) ^ -(n & 1);
-  }
-
-  /**
-   * Decode a ZigZag-encoded 64-bit value. ZigZag encodes signed integers into values that can be
-   * efficiently encoded with varint. (Otherwise, negative values must be sign-extended to 64 bits
-   * to be varint encoded, thus always taking 10 bytes on the wire.)
-   *
-   * @param n An unsigned 64-bit integer, stored in a signed int because Java has no explicit
-   *     unsigned support.
-   * @return A signed 64-bit integer.
-   */
-  public static long decodeZigZag64(final long n) {
-    return (n >>> 1) ^ -(n & 1);
-  }
-
-  /**
-   * Like {@link #readRawVarint32(InputStream)}, but expects that the caller has already read one
-   * byte. This allows the caller to determine if EOF has been reached before attempting to read.
-   */
-  public static int readRawVarint32(final int firstByte, final InputStream input)
-      throws IOException {
-    if ((firstByte & 0x80) == 0) {
-      return firstByte;
-    }
-
-    int result = firstByte & 0x7f;
-    int offset = 7;
-    for (; offset < 32; offset += 7) {
-      final int b = input.read();
-      if (b == -1) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-      result |= (b & 0x7f) << offset;
-      if ((b & 0x80) == 0) {
-        return result;
-      }
-    }
-    // Keep reading up to 64 bits.
-    for (; offset < 64; offset += 7) {
-      final int b = input.read();
-      if (b == -1) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-      if ((b & 0x80) == 0) {
-        return result;
-      }
-    }
-    throw InvalidProtocolBufferException.malformedVarint();
-  }
-
-  /**
-   * Reads a varint from the input one byte at a time, so that it does not read any bytes after the
-   * end of the varint. If you simply wrapped the stream in a CodedInputStream and used {@link
-   * #readRawVarint32(InputStream)} then you would probably end up reading past the end of the
-   * varint since CodedInputStream buffers its input.
-   */
-  static int readRawVarint32(final InputStream input) throws IOException {
-    final int firstByte = input.read();
-    if (firstByte == -1) {
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-    return readRawVarint32(firstByte, input);
-  }
-
-  /** A {@link CodedInputStream} implementation that uses a backing array as the input. */
-  private static final class ArrayDecoder extends CodedInputStream {
-    private final byte[] buffer;
-    private final boolean immutable;
-    private int limit;
-    private int bufferSizeAfterLimit;
-    private int pos;
-    private int startPos;
-    private int lastTag;
-    private boolean enableAliasing;
-
-    /** The absolute position of the end of the current message. */
-    private int currentLimit = Integer.MAX_VALUE;
-
-    private ArrayDecoder(final byte[] buffer, final int offset, final int len, boolean immutable) {
-      this.buffer = buffer;
-      limit = offset + len;
-      pos = offset;
-      startPos = pos;
-      this.immutable = immutable;
-    }
-
-    @Override
-    public int readTag() throws IOException {
-      if (isAtEnd()) {
-        lastTag = 0;
-        return 0;
-      }
-
-      lastTag = readRawVarint32();
-      if (WireFormat.getTagFieldNumber(lastTag) == 0) {
-        // If we actually read zero (or any tag number corresponding to field
-        // number zero), that's not a valid tag.
-        throw InvalidProtocolBufferException.invalidTag();
-      }
-      return lastTag;
-    }
-
-    @Override
-    public void checkLastTagWas(final int value) throws InvalidProtocolBufferException {
-      if (lastTag != value) {
-        throw InvalidProtocolBufferException.invalidEndTag();
-      }
-    }
-
-    @Override
-    public int getLastTag() {
-      return lastTag;
-    }
-
-    @Override
-    public boolean skipField(final int tag) throws IOException {
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          skipRawVarint();
-          return true;
-        case WireFormat.WIRETYPE_FIXED64:
-          skipRawBytes(FIXED_64_SIZE);
-          return true;
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          skipRawBytes(readRawVarint32());
-          return true;
-        case WireFormat.WIRETYPE_START_GROUP:
-          skipMessage();
-          checkLastTagWas(
-              WireFormat.makeTag(WireFormat.getTagFieldNumber(tag), WireFormat.WIRETYPE_END_GROUP));
-          return true;
-        case WireFormat.WIRETYPE_END_GROUP:
-          return false;
-        case WireFormat.WIRETYPE_FIXED32:
-          skipRawBytes(FIXED_32_SIZE);
-          return true;
-        default:
-          throw InvalidProtocolBufferException.invalidWireType();
-      }
-    }
-
-    @Override
-    public boolean skipField(final int tag, final CodedOutputStream output) throws IOException {
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          {
-            long value = readInt64();
-            output.writeRawVarint32(tag);
-            output.writeUInt64NoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_FIXED64:
-          {
-            long value = readRawLittleEndian64();
-            output.writeRawVarint32(tag);
-            output.writeFixed64NoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          {
-            ByteString value = readBytes();
-            output.writeRawVarint32(tag);
-            output.writeBytesNoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_START_GROUP:
-          {
-            output.writeRawVarint32(tag);
-            skipMessage(output);
-            int endtag =
-                WireFormat.makeTag(
-                    WireFormat.getTagFieldNumber(tag), WireFormat.WIRETYPE_END_GROUP);
-            checkLastTagWas(endtag);
-            output.writeRawVarint32(endtag);
-            return true;
-          }
-        case WireFormat.WIRETYPE_END_GROUP:
-          {
-            return false;
-          }
-        case WireFormat.WIRETYPE_FIXED32:
-          {
-            int value = readRawLittleEndian32();
-            output.writeRawVarint32(tag);
-            output.writeFixed32NoTag(value);
-            return true;
-          }
-        default:
-          throw InvalidProtocolBufferException.invalidWireType();
-      }
-    }
-
-    @Override
-    public void skipMessage() throws IOException {
-      while (true) {
-        final int tag = readTag();
-        if (tag == 0 || !skipField(tag)) {
-          return;
-        }
-      }
-    }
-
-    @Override
-    public void skipMessage(CodedOutputStream output) throws IOException {
-      while (true) {
-        final int tag = readTag();
-        if (tag == 0 || !skipField(tag, output)) {
-          return;
-        }
-      }
-    }
-
-
-    // -----------------------------------------------------------------
-
-    @Override
-    public double readDouble() throws IOException {
-      return Double.longBitsToDouble(readRawLittleEndian64());
-    }
-
-    @Override
-    public float readFloat() throws IOException {
-      return Float.intBitsToFloat(readRawLittleEndian32());
-    }
-
-    @Override
-    public long readUInt64() throws IOException {
-      return readRawVarint64();
-    }
-
-    @Override
-    public long readInt64() throws IOException {
-      return readRawVarint64();
-    }
-
-    @Override
-    public int readInt32() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public long readFixed64() throws IOException {
-      return readRawLittleEndian64();
-    }
-
-    @Override
-    public int readFixed32() throws IOException {
-      return readRawLittleEndian32();
-    }
-
-    @Override
-    public boolean readBool() throws IOException {
-      return readRawVarint64() != 0;
-    }
-
-    @Override
-    public String readString() throws IOException {
-      final int size = readRawVarint32();
-      if (size > 0 && size <= (limit - pos)) {
-        // Fast path:  We already have the bytes in a contiguous buffer, so
-        //   just copy directly from it.
-        final String result = new String(buffer, pos, size, UTF_8);
-        pos += size;
-        return result;
-      }
-
-      if (size == 0) {
-        return "";
-      }
-      if (size < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public String readStringRequireUtf8() throws IOException {
-      final int size = readRawVarint32();
-      if (size > 0 && size <= (limit - pos)) {
-        // TODO(martinrb): We could save a pass by validating while decoding.
-        if (!Utf8.isValidUtf8(buffer, pos, pos + size)) {
-          throw InvalidProtocolBufferException.invalidUtf8();
-        }
-        final int tempPos = pos;
-        pos += size;
-        return new String(buffer, tempPos, size, UTF_8);
-      }
-
-      if (size == 0) {
-        return "";
-      }
-      if (size <= 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public void readGroup(
-        final int fieldNumber,
-        final MessageLite.Builder builder,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      ++recursionDepth;
-      builder.mergeFrom(this, extensionRegistry);
-      checkLastTagWas(WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP));
-      --recursionDepth;
-    }
-
-
-    @Override
-    public <T extends MessageLite> T readGroup(
-        final int fieldNumber,
-        final Parser<T> parser,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      ++recursionDepth;
-      T result = parser.parsePartialFrom(this, extensionRegistry);
-      checkLastTagWas(WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP));
-      --recursionDepth;
-      return result;
-    }
-
-    @Deprecated
-    @Override
-    public void readUnknownGroup(final int fieldNumber, final MessageLite.Builder builder)
-        throws IOException {
-      readGroup(fieldNumber, builder, ExtensionRegistryLite.getEmptyRegistry());
-    }
-
-    @Override
-    public void readMessage(
-        final MessageLite.Builder builder, final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      final int length = readRawVarint32();
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      final int oldLimit = pushLimit(length);
-      ++recursionDepth;
-      builder.mergeFrom(this, extensionRegistry);
-      checkLastTagWas(0);
-      --recursionDepth;
-      popLimit(oldLimit);
-    }
-
-
-    @Override
-    public <T extends MessageLite> T readMessage(
-        final Parser<T> parser, final ExtensionRegistryLite extensionRegistry) throws IOException {
-      int length = readRawVarint32();
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      final int oldLimit = pushLimit(length);
-      ++recursionDepth;
-      T result = parser.parsePartialFrom(this, extensionRegistry);
-      checkLastTagWas(0);
-      --recursionDepth;
-      popLimit(oldLimit);
-      return result;
-    }
-
-    @Override
-    public ByteString readBytes() throws IOException {
-      final int size = readRawVarint32();
-      if (size > 0 && size <= (limit - pos)) {
-        // Fast path:  We already have the bytes in a contiguous buffer, so
-        //   just copy directly from it.
-        final ByteString result =
-            immutable && enableAliasing
-                ? ByteString.wrap(buffer, pos, size)
-                : ByteString.copyFrom(buffer, pos, size);
-        pos += size;
-        return result;
-      }
-      if (size == 0) {
-        return ByteString.EMPTY;
-      }
-      // Slow path:  Build a byte array first then copy it.
-      return ByteString.wrap(readRawBytes(size));
-    }
-
-    @Override
-    public byte[] readByteArray() throws IOException {
-      final int size = readRawVarint32();
-      return readRawBytes(size);
-    }
-
-    @Override
-    public ByteBuffer readByteBuffer() throws IOException {
-      final int size = readRawVarint32();
-      if (size > 0 && size <= (limit - pos)) {
-        // Fast path: We already have the bytes in a contiguous buffer.
-        // When aliasing is enabled, we can return a ByteBuffer pointing directly
-        // into the underlying byte array without copy if the CodedInputStream is
-        // constructed from a byte array. If aliasing is disabled or the input is
-        // from an InputStream or ByteString, we have to make a copy of the bytes.
-        ByteBuffer result =
-            !immutable && enableAliasing
-                ? ByteBuffer.wrap(buffer, pos, size).slice()
-                : ByteBuffer.wrap(Arrays.copyOfRange(buffer, pos, pos + size));
-        pos += size;
-        // TODO(nathanmittler): Investigate making the ByteBuffer be made read-only
-        return result;
-      }
-
-      if (size == 0) {
-        return EMPTY_BYTE_BUFFER;
-      }
-      if (size < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public int readUInt32() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public int readEnum() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public int readSFixed32() throws IOException {
-      return readRawLittleEndian32();
-    }
-
-    @Override
-    public long readSFixed64() throws IOException {
-      return readRawLittleEndian64();
-    }
-
-    @Override
-    public int readSInt32() throws IOException {
-      return decodeZigZag32(readRawVarint32());
-    }
-
-    @Override
-    public long readSInt64() throws IOException {
-      return decodeZigZag64(readRawVarint64());
-    }
-
-    // =================================================================
-
-    @Override
-    public int readRawVarint32() throws IOException {
-      // See implementation notes for readRawVarint64
-      fastpath:
-      {
-        int tempPos = pos;
-
-        if (limit == tempPos) {
-          break fastpath;
-        }
-
-        final byte[] buffer = this.buffer;
-        int x;
-        if ((x = buffer[tempPos++]) >= 0) {
-          pos = tempPos;
-          return x;
-        } else if (limit - tempPos < 9) {
-          break fastpath;
-        } else if ((x ^= (buffer[tempPos++] << 7)) < 0) {
-          x ^= (~0 << 7);
-        } else if ((x ^= (buffer[tempPos++] << 14)) >= 0) {
-          x ^= (~0 << 7) ^ (~0 << 14);
-        } else if ((x ^= (buffer[tempPos++] << 21)) < 0) {
-          x ^= (~0 << 7) ^ (~0 << 14) ^ (~0 << 21);
-        } else {
-          int y = buffer[tempPos++];
-          x ^= y << 28;
-          x ^= (~0 << 7) ^ (~0 << 14) ^ (~0 << 21) ^ (~0 << 28);
-          if (y < 0
-              && buffer[tempPos++] < 0
-              && buffer[tempPos++] < 0
-              && buffer[tempPos++] < 0
-              && buffer[tempPos++] < 0
-              && buffer[tempPos++] < 0) {
-            break fastpath; // Will throw malformedVarint()
-          }
-        }
-        pos = tempPos;
-        return x;
-      }
-      return (int) readRawVarint64SlowPath();
-    }
-
-    private void skipRawVarint() throws IOException {
-      if (limit - pos >= MAX_VARINT_SIZE) {
-        skipRawVarintFastPath();
-      } else {
-        skipRawVarintSlowPath();
-      }
-    }
-
-    private void skipRawVarintFastPath() throws IOException {
-      for (int i = 0; i < MAX_VARINT_SIZE; i++) {
-        if (buffer[pos++] >= 0) {
-          return;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    private void skipRawVarintSlowPath() throws IOException {
-      for (int i = 0; i < MAX_VARINT_SIZE; i++) {
-        if (readRawByte() >= 0) {
-          return;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    @Override
-    public long readRawVarint64() throws IOException {
-      // Implementation notes:
-      //
-      // Optimized for one-byte values, expected to be common.
-      // The particular code below was selected from various candidates
-      // empirically, by winning VarintBenchmark.
-      //
-      // Sign extension of (signed) Java bytes is usually a nuisance, but
-      // we exploit it here to more easily obtain the sign of bytes read.
-      // Instead of cleaning up the sign extension bits by masking eagerly,
-      // we delay until we find the final (positive) byte, when we clear all
-      // accumulated bits with one xor.  We depend on javac to constant fold.
-      fastpath:
-      {
-        int tempPos = pos;
-
-        if (limit == tempPos) {
-          break fastpath;
-        }
-
-        final byte[] buffer = this.buffer;
-        long x;
-        int y;
-        if ((y = buffer[tempPos++]) >= 0) {
-          pos = tempPos;
-          return y;
-        } else if (limit - tempPos < 9) {
-          break fastpath;
-        } else if ((y ^= (buffer[tempPos++] << 7)) < 0) {
-          x = y ^ (~0 << 7);
-        } else if ((y ^= (buffer[tempPos++] << 14)) >= 0) {
-          x = y ^ ((~0 << 7) ^ (~0 << 14));
-        } else if ((y ^= (buffer[tempPos++] << 21)) < 0) {
-          x = y ^ ((~0 << 7) ^ (~0 << 14) ^ (~0 << 21));
-        } else if ((x = y ^ ((long) buffer[tempPos++] << 28)) >= 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28);
-        } else if ((x ^= ((long) buffer[tempPos++] << 35)) < 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28) ^ (~0L << 35);
-        } else if ((x ^= ((long) buffer[tempPos++] << 42)) >= 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28) ^ (~0L << 35) ^ (~0L << 42);
-        } else if ((x ^= ((long) buffer[tempPos++] << 49)) < 0L) {
-          x ^=
-              (~0L << 7)
-                  ^ (~0L << 14)
-                  ^ (~0L << 21)
-                  ^ (~0L << 28)
-                  ^ (~0L << 35)
-                  ^ (~0L << 42)
-                  ^ (~0L << 49);
-        } else {
-          x ^= ((long) buffer[tempPos++] << 56);
-          x ^=
-              (~0L << 7)
-                  ^ (~0L << 14)
-                  ^ (~0L << 21)
-                  ^ (~0L << 28)
-                  ^ (~0L << 35)
-                  ^ (~0L << 42)
-                  ^ (~0L << 49)
-                  ^ (~0L << 56);
-          if (x < 0L) {
-            if (buffer[tempPos++] < 0L) {
-              break fastpath; // Will throw malformedVarint()
-            }
-          }
-        }
-        pos = tempPos;
-        return x;
-      }
-      return readRawVarint64SlowPath();
-    }
-
-    @Override
-    long readRawVarint64SlowPath() throws IOException {
-      long result = 0;
-      for (int shift = 0; shift < 64; shift += 7) {
-        final byte b = readRawByte();
-        result |= (long) (b & 0x7F) << shift;
-        if ((b & 0x80) == 0) {
-          return result;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    @Override
-    public int readRawLittleEndian32() throws IOException {
-      int tempPos = pos;
-
-      if (limit - tempPos < FIXED_32_SIZE) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-
-      final byte[] buffer = this.buffer;
-      pos = tempPos + FIXED_32_SIZE;
-      return (((buffer[tempPos] & 0xff))
-          | ((buffer[tempPos + 1] & 0xff) << 8)
-          | ((buffer[tempPos + 2] & 0xff) << 16)
-          | ((buffer[tempPos + 3] & 0xff) << 24));
-    }
-
-    @Override
-    public long readRawLittleEndian64() throws IOException {
-      int tempPos = pos;
-
-      if (limit - tempPos < FIXED_64_SIZE) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-
-      final byte[] buffer = this.buffer;
-      pos = tempPos + FIXED_64_SIZE;
-      return (((buffer[tempPos] & 0xffL))
-          | ((buffer[tempPos + 1] & 0xffL) << 8)
-          | ((buffer[tempPos + 2] & 0xffL) << 16)
-          | ((buffer[tempPos + 3] & 0xffL) << 24)
-          | ((buffer[tempPos + 4] & 0xffL) << 32)
-          | ((buffer[tempPos + 5] & 0xffL) << 40)
-          | ((buffer[tempPos + 6] & 0xffL) << 48)
-          | ((buffer[tempPos + 7] & 0xffL) << 56));
-    }
-
-    @Override
-    public void enableAliasing(boolean enabled) {
-      this.enableAliasing = enabled;
-    }
-
-    @Override
-    public void resetSizeCounter() {
-      startPos = pos;
-    }
-
-    @Override
-    public int pushLimit(int byteLimit) throws InvalidProtocolBufferException {
-      if (byteLimit < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      byteLimit += getTotalBytesRead();
-      final int oldLimit = currentLimit;
-      if (byteLimit > oldLimit) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-      currentLimit = byteLimit;
-
-      recomputeBufferSizeAfterLimit();
-
-      return oldLimit;
-    }
-
-    private void recomputeBufferSizeAfterLimit() {
-      limit += bufferSizeAfterLimit;
-      final int bufferEnd = limit - startPos;
-      if (bufferEnd > currentLimit) {
-        // Limit is in current buffer.
-        bufferSizeAfterLimit = bufferEnd - currentLimit;
-        limit -= bufferSizeAfterLimit;
-      } else {
-        bufferSizeAfterLimit = 0;
-      }
-    }
-
-    @Override
-    public void popLimit(final int oldLimit) {
-      currentLimit = oldLimit;
-      recomputeBufferSizeAfterLimit();
-    }
-
-    @Override
-    public int getBytesUntilLimit() {
-      if (currentLimit == Integer.MAX_VALUE) {
-        return -1;
-      }
-
-      return currentLimit - getTotalBytesRead();
-    }
-
-    @Override
-    public boolean isAtEnd() throws IOException {
-      return pos == limit;
-    }
-
-    @Override
-    public int getTotalBytesRead() {
-      return pos - startPos;
-    }
-
-    @Override
-    public byte readRawByte() throws IOException {
-      if (pos == limit) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-      return buffer[pos++];
-    }
-
-    @Override
-    public byte[] readRawBytes(final int length) throws IOException {
-      if (length > 0 && length <= (limit - pos)) {
-        final int tempPos = pos;
-        pos += length;
-        return Arrays.copyOfRange(buffer, tempPos, pos);
-      }
-
-      if (length <= 0) {
-        if (length == 0) {
-          return Internal.EMPTY_BYTE_ARRAY;
-        } else {
-          throw InvalidProtocolBufferException.negativeSize();
-        }
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public void skipRawBytes(final int length) throws IOException {
-      if (length >= 0 && length <= (limit - pos)) {
-        // We have all the bytes we need already.
-        pos += length;
-        return;
-      }
-
-      if (length < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-  }
-
-  /**
-   * A {@link CodedInputStream} implementation that uses a backing direct ByteBuffer as the input.
-   * Requires the use of {@code sun.misc.Unsafe} to perform fast reads on the buffer.
-   */
-  private static final class UnsafeDirectNioDecoder extends CodedInputStream {
-    /** The direct buffer that is backing this stream. */
-    private final ByteBuffer buffer;
-
-    /**
-     * If {@code true}, indicates that the buffer is backing a {@link ByteString} and is therefore
-     * considered to be an immutable input source.
-     */
-    private final boolean immutable;
-
-    /** The unsafe address of the content of {@link #buffer}. */
-    private final long address;
-
-    /** The unsafe address of the current read limit of the buffer. */
-    private long limit;
-
-    /** The unsafe address of the current read position of the buffer. */
-    private long pos;
-
-    /** The unsafe address of the starting read position. */
-    private long startPos;
-
-    /** The amount of available data in the buffer beyond {@link #limit}. */
-    private int bufferSizeAfterLimit;
-
-    /** The last tag that was read from this stream. */
-    private int lastTag;
-
-    /**
-     * If {@code true}, indicates that calls to read {@link ByteString} or {@code byte[]}
-     * <strong>may</strong> return slices of the underlying buffer, rather than copies.
-     */
-    private boolean enableAliasing;
-
-    /** The absolute position of the end of the current message. */
-    private int currentLimit = Integer.MAX_VALUE;
-
-    static boolean isSupported() {
-      return UnsafeUtil.hasUnsafeByteBufferOperations();
-    }
-
-    private UnsafeDirectNioDecoder(ByteBuffer buffer, boolean immutable) {
-      this.buffer = buffer;
-      address = UnsafeUtil.addressOffset(buffer);
-      limit = address + buffer.limit();
-      pos = address + buffer.position();
-      startPos = pos;
-      this.immutable = immutable;
-    }
-
-    @Override
-    public int readTag() throws IOException {
-      if (isAtEnd()) {
-        lastTag = 0;
-        return 0;
-      }
-
-      lastTag = readRawVarint32();
-      if (WireFormat.getTagFieldNumber(lastTag) == 0) {
-        // If we actually read zero (or any tag number corresponding to field
-        // number zero), that's not a valid tag.
-        throw InvalidProtocolBufferException.invalidTag();
-      }
-      return lastTag;
-    }
-
-    @Override
-    public void checkLastTagWas(final int value) throws InvalidProtocolBufferException {
-      if (lastTag != value) {
-        throw InvalidProtocolBufferException.invalidEndTag();
-      }
-    }
-
-    @Override
-    public int getLastTag() {
-      return lastTag;
-    }
-
-    @Override
-    public boolean skipField(final int tag) throws IOException {
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          skipRawVarint();
-          return true;
-        case WireFormat.WIRETYPE_FIXED64:
-          skipRawBytes(FIXED_64_SIZE);
-          return true;
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          skipRawBytes(readRawVarint32());
-          return true;
-        case WireFormat.WIRETYPE_START_GROUP:
-          skipMessage();
-          checkLastTagWas(
-              WireFormat.makeTag(WireFormat.getTagFieldNumber(tag), WireFormat.WIRETYPE_END_GROUP));
-          return true;
-        case WireFormat.WIRETYPE_END_GROUP:
-          return false;
-        case WireFormat.WIRETYPE_FIXED32:
-          skipRawBytes(FIXED_32_SIZE);
-          return true;
-        default:
-          throw InvalidProtocolBufferException.invalidWireType();
-      }
-    }
-
-    @Override
-    public boolean skipField(final int tag, final CodedOutputStream output) throws IOException {
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          {
-            long value = readInt64();
-            output.writeRawVarint32(tag);
-            output.writeUInt64NoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_FIXED64:
-          {
-            long value = readRawLittleEndian64();
-            output.writeRawVarint32(tag);
-            output.writeFixed64NoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          {
-            ByteString value = readBytes();
-            output.writeRawVarint32(tag);
-            output.writeBytesNoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_START_GROUP:
-          {
-            output.writeRawVarint32(tag);
-            skipMessage(output);
-            int endtag =
-                WireFormat.makeTag(
-                    WireFormat.getTagFieldNumber(tag), WireFormat.WIRETYPE_END_GROUP);
-            checkLastTagWas(endtag);
-            output.writeRawVarint32(endtag);
-            return true;
-          }
-        case WireFormat.WIRETYPE_END_GROUP:
-          {
-            return false;
-          }
-        case WireFormat.WIRETYPE_FIXED32:
-          {
-            int value = readRawLittleEndian32();
-            output.writeRawVarint32(tag);
-            output.writeFixed32NoTag(value);
-            return true;
-          }
-        default:
-          throw InvalidProtocolBufferException.invalidWireType();
-      }
-    }
-
-    @Override
-    public void skipMessage() throws IOException {
-      while (true) {
-        final int tag = readTag();
-        if (tag == 0 || !skipField(tag)) {
-          return;
-        }
-      }
-    }
-
-    @Override
-    public void skipMessage(CodedOutputStream output) throws IOException {
-      while (true) {
-        final int tag = readTag();
-        if (tag == 0 || !skipField(tag, output)) {
-          return;
-        }
-      }
-    }
-
-
-    // -----------------------------------------------------------------
-
-    @Override
-    public double readDouble() throws IOException {
-      return Double.longBitsToDouble(readRawLittleEndian64());
-    }
-
-    @Override
-    public float readFloat() throws IOException {
-      return Float.intBitsToFloat(readRawLittleEndian32());
-    }
-
-    @Override
-    public long readUInt64() throws IOException {
-      return readRawVarint64();
-    }
-
-    @Override
-    public long readInt64() throws IOException {
-      return readRawVarint64();
-    }
-
-    @Override
-    public int readInt32() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public long readFixed64() throws IOException {
-      return readRawLittleEndian64();
-    }
-
-    @Override
-    public int readFixed32() throws IOException {
-      return readRawLittleEndian32();
-    }
-
-    @Override
-    public boolean readBool() throws IOException {
-      return readRawVarint64() != 0;
-    }
-
-    @Override
-    public String readString() throws IOException {
-      final int size = readRawVarint32();
-      if (size > 0 && size <= remaining()) {
-        // TODO(nathanmittler): Is there a way to avoid this copy?
-        byte[] bytes = copyToArray(pos, pos + size);
-        String result = new String(bytes, UTF_8);
-        pos += size;
-        return result;
-      }
-
-      if (size == 0) {
-        return "";
-      }
-      if (size < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public String readStringRequireUtf8() throws IOException {
-      final int size = readRawVarint32();
-      if (size >= 0 && size <= remaining()) {
-        // TODO(nathanmittler): Is there a way to avoid this copy?
-        byte[] bytes = copyToArray(pos, pos + size);
-        // TODO(martinrb): We could save a pass by validating while decoding.
-        if (!Utf8.isValidUtf8(bytes)) {
-          throw InvalidProtocolBufferException.invalidUtf8();
-        }
-
-        String result = new String(bytes, UTF_8);
-        pos += size;
-        return result;
-      }
-
-      if (size == 0) {
-        return "";
-      }
-      if (size <= 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public void readGroup(
-        final int fieldNumber,
-        final MessageLite.Builder builder,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      ++recursionDepth;
-      builder.mergeFrom(this, extensionRegistry);
-      checkLastTagWas(WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP));
-      --recursionDepth;
-    }
-
-
-    @Override
-    public <T extends MessageLite> T readGroup(
-        final int fieldNumber,
-        final Parser<T> parser,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      ++recursionDepth;
-      T result = parser.parsePartialFrom(this, extensionRegistry);
-      checkLastTagWas(WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP));
-      --recursionDepth;
-      return result;
-    }
-
-    @Deprecated
-    @Override
-    public void readUnknownGroup(final int fieldNumber, final MessageLite.Builder builder)
-        throws IOException {
-      readGroup(fieldNumber, builder, ExtensionRegistryLite.getEmptyRegistry());
-    }
-
-    @Override
-    public void readMessage(
-        final MessageLite.Builder builder, final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      final int length = readRawVarint32();
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      final int oldLimit = pushLimit(length);
-      ++recursionDepth;
-      builder.mergeFrom(this, extensionRegistry);
-      checkLastTagWas(0);
-      --recursionDepth;
-      popLimit(oldLimit);
-    }
-
-
-    @Override
-    public <T extends MessageLite> T readMessage(
-        final Parser<T> parser, final ExtensionRegistryLite extensionRegistry) throws IOException {
-      int length = readRawVarint32();
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      final int oldLimit = pushLimit(length);
-      ++recursionDepth;
-      T result = parser.parsePartialFrom(this, extensionRegistry);
-      checkLastTagWas(0);
-      --recursionDepth;
-      popLimit(oldLimit);
-      return result;
-    }
-
-    @Override
-    public ByteString readBytes() throws IOException {
-      final int size = readRawVarint32();
-      if (size > 0 && size <= remaining()) {
-        ByteBuffer result;
-        if (immutable && enableAliasing) {
-          result = slice(pos, pos + size);
-        } else {
-          result = copy(pos, pos + size);
-        }
-        pos += size;
-        return ByteString.wrap(result);
-      }
-
-      if (size == 0) {
-        return ByteString.EMPTY;
-      }
-      if (size < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public byte[] readByteArray() throws IOException {
-      return readRawBytes(readRawVarint32());
-    }
-
-    @Override
-    public ByteBuffer readByteBuffer() throws IOException {
-      final int size = readRawVarint32();
-      if (size > 0 && size <= remaining()) {
-        ByteBuffer result;
-        // "Immutable" implies that buffer is backing a ByteString.
-        // Disallow slicing in this case to prevent the caller from modifying the contents
-        // of the ByteString.
-        if (!immutable && enableAliasing) {
-          result = slice(pos, pos + size);
-        } else {
-          result = copy(pos, pos + size);
-        }
-        pos += size;
-        // TODO(nathanmittler): Investigate making the ByteBuffer be made read-only
-        return result;
-      }
-
-      if (size == 0) {
-        return EMPTY_BYTE_BUFFER;
-      }
-      if (size < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public int readUInt32() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public int readEnum() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public int readSFixed32() throws IOException {
-      return readRawLittleEndian32();
-    }
-
-    @Override
-    public long readSFixed64() throws IOException {
-      return readRawLittleEndian64();
-    }
-
-    @Override
-    public int readSInt32() throws IOException {
-      return decodeZigZag32(readRawVarint32());
-    }
-
-    @Override
-    public long readSInt64() throws IOException {
-      return decodeZigZag64(readRawVarint64());
-    }
-
-    // =================================================================
-
-    @Override
-    public int readRawVarint32() throws IOException {
-      // See implementation notes for readRawVarint64
-      fastpath:
-      {
-        long tempPos = pos;
-
-        if (limit == tempPos) {
-          break fastpath;
-        }
-
-        int x;
-        if ((x = UnsafeUtil.getByte(tempPos++)) >= 0) {
-          pos = tempPos;
-          return x;
-        } else if (limit - tempPos < 9) {
-          break fastpath;
-        } else if ((x ^= (UnsafeUtil.getByte(tempPos++) << 7)) < 0) {
-          x ^= (~0 << 7);
-        } else if ((x ^= (UnsafeUtil.getByte(tempPos++) << 14)) >= 0) {
-          x ^= (~0 << 7) ^ (~0 << 14);
-        } else if ((x ^= (UnsafeUtil.getByte(tempPos++) << 21)) < 0) {
-          x ^= (~0 << 7) ^ (~0 << 14) ^ (~0 << 21);
-        } else {
-          int y = UnsafeUtil.getByte(tempPos++);
-          x ^= y << 28;
-          x ^= (~0 << 7) ^ (~0 << 14) ^ (~0 << 21) ^ (~0 << 28);
-          if (y < 0
-              && UnsafeUtil.getByte(tempPos++) < 0
-              && UnsafeUtil.getByte(tempPos++) < 0
-              && UnsafeUtil.getByte(tempPos++) < 0
-              && UnsafeUtil.getByte(tempPos++) < 0
-              && UnsafeUtil.getByte(tempPos++) < 0) {
-            break fastpath; // Will throw malformedVarint()
-          }
-        }
-        pos = tempPos;
-        return x;
-      }
-      return (int) readRawVarint64SlowPath();
-    }
-
-    private void skipRawVarint() throws IOException {
-      if (remaining() >= MAX_VARINT_SIZE) {
-        skipRawVarintFastPath();
-      } else {
-        skipRawVarintSlowPath();
-      }
-    }
-
-    private void skipRawVarintFastPath() throws IOException {
-      for (int i = 0; i < MAX_VARINT_SIZE; i++) {
-        if (UnsafeUtil.getByte(pos++) >= 0) {
-          return;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    private void skipRawVarintSlowPath() throws IOException {
-      for (int i = 0; i < MAX_VARINT_SIZE; i++) {
-        if (readRawByte() >= 0) {
-          return;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    @Override
-    public long readRawVarint64() throws IOException {
-      // Implementation notes:
-      //
-      // Optimized for one-byte values, expected to be common.
-      // The particular code below was selected from various candidates
-      // empirically, by winning VarintBenchmark.
-      //
-      // Sign extension of (signed) Java bytes is usually a nuisance, but
-      // we exploit it here to more easily obtain the sign of bytes read.
-      // Instead of cleaning up the sign extension bits by masking eagerly,
-      // we delay until we find the final (positive) byte, when we clear all
-      // accumulated bits with one xor.  We depend on javac to constant fold.
-      fastpath:
-      {
-        long tempPos = pos;
-
-        if (limit == tempPos) {
-          break fastpath;
-        }
-
-        long x;
-        int y;
-        if ((y = UnsafeUtil.getByte(tempPos++)) >= 0) {
-          pos = tempPos;
-          return y;
-        } else if (limit - tempPos < 9) {
-          break fastpath;
-        } else if ((y ^= (UnsafeUtil.getByte(tempPos++) << 7)) < 0) {
-          x = y ^ (~0 << 7);
-        } else if ((y ^= (UnsafeUtil.getByte(tempPos++) << 14)) >= 0) {
-          x = y ^ ((~0 << 7) ^ (~0 << 14));
-        } else if ((y ^= (UnsafeUtil.getByte(tempPos++) << 21)) < 0) {
-          x = y ^ ((~0 << 7) ^ (~0 << 14) ^ (~0 << 21));
-        } else if ((x = y ^ ((long) UnsafeUtil.getByte(tempPos++) << 28)) >= 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28);
-        } else if ((x ^= ((long) UnsafeUtil.getByte(tempPos++) << 35)) < 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28) ^ (~0L << 35);
-        } else if ((x ^= ((long) UnsafeUtil.getByte(tempPos++) << 42)) >= 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28) ^ (~0L << 35) ^ (~0L << 42);
-        } else if ((x ^= ((long) UnsafeUtil.getByte(tempPos++) << 49)) < 0L) {
-          x ^=
-              (~0L << 7)
-                  ^ (~0L << 14)
-                  ^ (~0L << 21)
-                  ^ (~0L << 28)
-                  ^ (~0L << 35)
-                  ^ (~0L << 42)
-                  ^ (~0L << 49);
-        } else {
-          x ^= ((long) UnsafeUtil.getByte(tempPos++) << 56);
-          x ^=
-              (~0L << 7)
-                  ^ (~0L << 14)
-                  ^ (~0L << 21)
-                  ^ (~0L << 28)
-                  ^ (~0L << 35)
-                  ^ (~0L << 42)
-                  ^ (~0L << 49)
-                  ^ (~0L << 56);
-          if (x < 0L) {
-            if (UnsafeUtil.getByte(tempPos++) < 0L) {
-              break fastpath; // Will throw malformedVarint()
-            }
-          }
-        }
-        pos = tempPos;
-        return x;
-      }
-      return readRawVarint64SlowPath();
-    }
-
-    @Override
-    long readRawVarint64SlowPath() throws IOException {
-      long result = 0;
-      for (int shift = 0; shift < 64; shift += 7) {
-        final byte b = readRawByte();
-        result |= (long) (b & 0x7F) << shift;
-        if ((b & 0x80) == 0) {
-          return result;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    @Override
-    public int readRawLittleEndian32() throws IOException {
-      long tempPos = pos;
-
-      if (limit - tempPos < FIXED_32_SIZE) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-
-      pos = tempPos + FIXED_32_SIZE;
-      return (((UnsafeUtil.getByte(tempPos) & 0xff))
-          | ((UnsafeUtil.getByte(tempPos + 1) & 0xff) << 8)
-          | ((UnsafeUtil.getByte(tempPos + 2) & 0xff) << 16)
-          | ((UnsafeUtil.getByte(tempPos + 3) & 0xff) << 24));
-    }
-
-    @Override
-    public long readRawLittleEndian64() throws IOException {
-      long tempPos = pos;
-
-      if (limit - tempPos < FIXED_64_SIZE) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-
-      pos = tempPos + FIXED_64_SIZE;
-      return (((UnsafeUtil.getByte(tempPos) & 0xffL))
-          | ((UnsafeUtil.getByte(tempPos + 1) & 0xffL) << 8)
-          | ((UnsafeUtil.getByte(tempPos + 2) & 0xffL) << 16)
-          | ((UnsafeUtil.getByte(tempPos + 3) & 0xffL) << 24)
-          | ((UnsafeUtil.getByte(tempPos + 4) & 0xffL) << 32)
-          | ((UnsafeUtil.getByte(tempPos + 5) & 0xffL) << 40)
-          | ((UnsafeUtil.getByte(tempPos + 6) & 0xffL) << 48)
-          | ((UnsafeUtil.getByte(tempPos + 7) & 0xffL) << 56));
-    }
-
-    @Override
-    public void enableAliasing(boolean enabled) {
-      this.enableAliasing = enabled;
-    }
-
-    @Override
-    public void resetSizeCounter() {
-      startPos = pos;
-    }
-
-    @Override
-    public int pushLimit(int byteLimit) throws InvalidProtocolBufferException {
-      if (byteLimit < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      byteLimit += getTotalBytesRead();
-      final int oldLimit = currentLimit;
-      if (byteLimit > oldLimit) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-      currentLimit = byteLimit;
-
-      recomputeBufferSizeAfterLimit();
-
-      return oldLimit;
-    }
-
-    @Override
-    public void popLimit(final int oldLimit) {
-      currentLimit = oldLimit;
-      recomputeBufferSizeAfterLimit();
-    }
-
-    @Override
-    public int getBytesUntilLimit() {
-      if (currentLimit == Integer.MAX_VALUE) {
-        return -1;
-      }
-
-      return currentLimit - getTotalBytesRead();
-    }
-
-    @Override
-    public boolean isAtEnd() throws IOException {
-      return pos == limit;
-    }
-
-    @Override
-    public int getTotalBytesRead() {
-      return (int) (pos - startPos);
-    }
-
-    @Override
-    public byte readRawByte() throws IOException {
-      if (pos == limit) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-      return UnsafeUtil.getByte(pos++);
-    }
-
-    @Override
-    public byte[] readRawBytes(final int length) throws IOException {
-      if (length >= 0 && length <= remaining()) {
-        byte[] bytes = new byte[length];
-        slice(pos, pos + length).get(bytes);
-        pos += length;
-        return bytes;
-      }
-
-      if (length <= 0) {
-        if (length == 0) {
-          return EMPTY_BYTE_ARRAY;
-        } else {
-          throw InvalidProtocolBufferException.negativeSize();
-        }
-      }
-
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    @Override
-    public void skipRawBytes(final int length) throws IOException {
-      if (length >= 0 && length <= remaining()) {
-        // We have all the bytes we need already.
-        pos += length;
-        return;
-      }
-
-      if (length < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      throw InvalidProtocolBufferException.truncatedMessage();
-    }
-
-    private void recomputeBufferSizeAfterLimit() {
-      limit += bufferSizeAfterLimit;
-      final int bufferEnd = (int) (limit - startPos);
-      if (bufferEnd > currentLimit) {
-        // Limit is in current buffer.
-        bufferSizeAfterLimit = bufferEnd - currentLimit;
-        limit -= bufferSizeAfterLimit;
-      } else {
-        bufferSizeAfterLimit = 0;
-      }
-    }
-
-    private int remaining() {
-      return (int) (limit - pos);
-    }
-
-    private int bufferPos(long pos) {
-      return (int) (pos - address);
-    }
-
-    private ByteBuffer slice(long begin, long end) throws IOException {
-      int prevPos = buffer.position();
-      int prevLimit = buffer.limit();
-      try {
-        buffer.position(bufferPos(begin));
-        buffer.limit(bufferPos(end));
-        return buffer.slice();
-      } catch (IllegalArgumentException e) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      } finally {
-        buffer.position(prevPos);
-        buffer.limit(prevLimit);
-      }
-    }
-
-    private ByteBuffer copy(long begin, long end) throws IOException {
-      return ByteBuffer.wrap(copyToArray(begin, end));
-    }
-
-    private byte[] copyToArray(long begin, long end) throws IOException {
-      int prevPos = buffer.position();
-      int prevLimit = buffer.limit();
-      try {
-        buffer.position(bufferPos(begin));
-        buffer.limit(bufferPos(end));
-        byte[] bytes = new byte[(int) (end - begin)];
-        buffer.get(bytes);
-        return bytes;
-      } catch (IllegalArgumentException e) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      } finally {
-        buffer.position(prevPos);
-        buffer.limit(prevLimit);
-      }
-    }
-  }
-
-  /**
-   * Implementation of {@link CodedInputStream} that uses an {@link InputStream} as the data source.
-   */
-  private static final class StreamDecoder extends CodedInputStream {
-    private final InputStream input;
-    private final byte[] buffer;
-    /** bufferSize represents how many bytes are currently filled in the buffer */
-    private int bufferSize;
-
-    private int bufferSizeAfterLimit;
-    private int pos;
-    private int lastTag;
-
-    /**
-     * The total number of bytes read before the current buffer. The total bytes read up to the
-     * current position can be computed as {@code totalBytesRetired + pos}. This value may be
-     * negative if reading started in the middle of the current buffer (e.g. if the constructor that
-     * takes a byte array and an offset was used).
-     */
-    private int totalBytesRetired;
-
-    /** The absolute position of the end of the current message. */
-    private int currentLimit = Integer.MAX_VALUE;
-
-    private StreamDecoder(final InputStream input, int bufferSize) {
-      checkNotNull(input, "input");
-      this.input = input;
-      this.buffer = new byte[bufferSize];
-      this.bufferSize = 0;
-      pos = 0;
-      totalBytesRetired = 0;
-    }
-
-    @Override
-    public int readTag() throws IOException {
-      if (isAtEnd()) {
-        lastTag = 0;
-        return 0;
-      }
-
-      lastTag = readRawVarint32();
-      if (WireFormat.getTagFieldNumber(lastTag) == 0) {
-        // If we actually read zero (or any tag number corresponding to field
-        // number zero), that's not a valid tag.
-        throw InvalidProtocolBufferException.invalidTag();
-      }
-      return lastTag;
-    }
-
-    @Override
-    public void checkLastTagWas(final int value) throws InvalidProtocolBufferException {
-      if (lastTag != value) {
-        throw InvalidProtocolBufferException.invalidEndTag();
-      }
-    }
-
-    @Override
-    public int getLastTag() {
-      return lastTag;
-    }
-
-    @Override
-    public boolean skipField(final int tag) throws IOException {
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          skipRawVarint();
-          return true;
-        case WireFormat.WIRETYPE_FIXED64:
-          skipRawBytes(FIXED_64_SIZE);
-          return true;
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          skipRawBytes(readRawVarint32());
-          return true;
-        case WireFormat.WIRETYPE_START_GROUP:
-          skipMessage();
-          checkLastTagWas(
-              WireFormat.makeTag(WireFormat.getTagFieldNumber(tag), WireFormat.WIRETYPE_END_GROUP));
-          return true;
-        case WireFormat.WIRETYPE_END_GROUP:
-          return false;
-        case WireFormat.WIRETYPE_FIXED32:
-          skipRawBytes(FIXED_32_SIZE);
-          return true;
-        default:
-          throw InvalidProtocolBufferException.invalidWireType();
-      }
-    }
-
-    @Override
-    public boolean skipField(final int tag, final CodedOutputStream output) throws IOException {
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          {
-            long value = readInt64();
-            output.writeRawVarint32(tag);
-            output.writeUInt64NoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_FIXED64:
-          {
-            long value = readRawLittleEndian64();
-            output.writeRawVarint32(tag);
-            output.writeFixed64NoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          {
-            ByteString value = readBytes();
-            output.writeRawVarint32(tag);
-            output.writeBytesNoTag(value);
-            return true;
-          }
-        case WireFormat.WIRETYPE_START_GROUP:
-          {
-            output.writeRawVarint32(tag);
-            skipMessage(output);
-            int endtag =
-                WireFormat.makeTag(
-                    WireFormat.getTagFieldNumber(tag), WireFormat.WIRETYPE_END_GROUP);
-            checkLastTagWas(endtag);
-            output.writeRawVarint32(endtag);
-            return true;
-          }
-        case WireFormat.WIRETYPE_END_GROUP:
-          {
-            return false;
-          }
-        case WireFormat.WIRETYPE_FIXED32:
-          {
-            int value = readRawLittleEndian32();
-            output.writeRawVarint32(tag);
-            output.writeFixed32NoTag(value);
-            return true;
-          }
-        default:
-          throw InvalidProtocolBufferException.invalidWireType();
-      }
-    }
-
-    @Override
-    public void skipMessage() throws IOException {
-      while (true) {
-        final int tag = readTag();
-        if (tag == 0 || !skipField(tag)) {
-          return;
-        }
-      }
-    }
-
-    @Override
-    public void skipMessage(CodedOutputStream output) throws IOException {
-      while (true) {
-        final int tag = readTag();
-        if (tag == 0 || !skipField(tag, output)) {
-          return;
-        }
-      }
-    }
-
-    /** Collects the bytes skipped and returns the data in a ByteBuffer. */
-    private class SkippedDataSink implements RefillCallback {
-      private int lastPos = pos;
-      private ByteArrayOutputStream byteArrayStream;
-
-      @Override
-      public void onRefill() {
-        if (byteArrayStream == null) {
-          byteArrayStream = new ByteArrayOutputStream();
-        }
-        byteArrayStream.write(buffer, lastPos, pos - lastPos);
-        lastPos = 0;
-      }
-
-      /** Gets skipped data in a ByteBuffer. This method should only be called once. */
-      ByteBuffer getSkippedData() {
-        if (byteArrayStream == null) {
-          return ByteBuffer.wrap(buffer, lastPos, pos - lastPos);
-        } else {
-          byteArrayStream.write(buffer, lastPos, pos);
-          return ByteBuffer.wrap(byteArrayStream.toByteArray());
-        }
-      }
-    }
-
-
-    // -----------------------------------------------------------------
-
-    @Override
-    public double readDouble() throws IOException {
-      return Double.longBitsToDouble(readRawLittleEndian64());
-    }
-
-    @Override
-    public float readFloat() throws IOException {
-      return Float.intBitsToFloat(readRawLittleEndian32());
-    }
-
-    @Override
-    public long readUInt64() throws IOException {
-      return readRawVarint64();
-    }
-
-    @Override
-    public long readInt64() throws IOException {
-      return readRawVarint64();
-    }
-
-    @Override
-    public int readInt32() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public long readFixed64() throws IOException {
-      return readRawLittleEndian64();
-    }
-
-    @Override
-    public int readFixed32() throws IOException {
-      return readRawLittleEndian32();
-    }
-
-    @Override
-    public boolean readBool() throws IOException {
-      return readRawVarint64() != 0;
-    }
-
-    @Override
-    public String readString() throws IOException {
-      final int size = readRawVarint32();
-      if (size > 0 && size <= (bufferSize - pos)) {
-        // Fast path:  We already have the bytes in a contiguous buffer, so
-        //   just copy directly from it.
-        final String result = new String(buffer, pos, size, UTF_8);
-        pos += size;
-        return result;
-      }
-      if (size == 0) {
-        return "";
-      }
-      if (size <= bufferSize) {
-        refillBuffer(size);
-        String result = new String(buffer, pos, size, UTF_8);
-        pos += size;
-        return result;
-      }
-      // Slow path:  Build a byte array first then copy it.
-      return new String(readRawBytesSlowPath(size), UTF_8);
-    }
-
-    @Override
-    public String readStringRequireUtf8() throws IOException {
-      final int size = readRawVarint32();
-      final byte[] bytes;
-      final int oldPos = pos;
-      final int tempPos;
-      if (size <= (bufferSize - oldPos) && size > 0) {
-        // Fast path:  We already have the bytes in a contiguous buffer, so
-        //   just copy directly from it.
-        bytes = buffer;
-        pos = oldPos + size;
-        tempPos = oldPos;
-      } else if (size == 0) {
-        return "";
-      } else if (size <= bufferSize) {
-        refillBuffer(size);
-        bytes = buffer;
-        tempPos = 0;
-        pos = tempPos + size;
-      } else {
-        // Slow path:  Build a byte array first then copy it.
-        bytes = readRawBytesSlowPath(size);
-        tempPos = 0;
-      }
-      // TODO(martinrb): We could save a pass by validating while decoding.
-      if (!Utf8.isValidUtf8(bytes, tempPos, tempPos + size)) {
-        throw InvalidProtocolBufferException.invalidUtf8();
-      }
-      return new String(bytes, tempPos, size, UTF_8);
-    }
-
-    @Override
-    public void readGroup(
-        final int fieldNumber,
-        final MessageLite.Builder builder,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      ++recursionDepth;
-      builder.mergeFrom(this, extensionRegistry);
-      checkLastTagWas(WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP));
-      --recursionDepth;
-    }
-
-
-    @Override
-    public <T extends MessageLite> T readGroup(
-        final int fieldNumber,
-        final Parser<T> parser,
-        final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      ++recursionDepth;
-      T result = parser.parsePartialFrom(this, extensionRegistry);
-      checkLastTagWas(WireFormat.makeTag(fieldNumber, WireFormat.WIRETYPE_END_GROUP));
-      --recursionDepth;
-      return result;
-    }
-
-    @Deprecated
-    @Override
-    public void readUnknownGroup(final int fieldNumber, final MessageLite.Builder builder)
-        throws IOException {
-      readGroup(fieldNumber, builder, ExtensionRegistryLite.getEmptyRegistry());
-    }
-
-    @Override
-    public void readMessage(
-        final MessageLite.Builder builder, final ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      final int length = readRawVarint32();
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      final int oldLimit = pushLimit(length);
-      ++recursionDepth;
-      builder.mergeFrom(this, extensionRegistry);
-      checkLastTagWas(0);
-      --recursionDepth;
-      popLimit(oldLimit);
-    }
-
-
-    @Override
-    public <T extends MessageLite> T readMessage(
-        final Parser<T> parser, final ExtensionRegistryLite extensionRegistry) throws IOException {
-      int length = readRawVarint32();
-      if (recursionDepth >= recursionLimit) {
-        throw InvalidProtocolBufferException.recursionLimitExceeded();
-      }
-      final int oldLimit = pushLimit(length);
-      ++recursionDepth;
-      T result = parser.parsePartialFrom(this, extensionRegistry);
-      checkLastTagWas(0);
-      --recursionDepth;
-      popLimit(oldLimit);
-      return result;
-    }
-
-    @Override
-    public ByteString readBytes() throws IOException {
-      final int size = readRawVarint32();
-      if (size <= (bufferSize - pos) && size > 0) {
-        // Fast path:  We already have the bytes in a contiguous buffer, so
-        //   just copy directly from it.
-        final ByteString result = ByteString.copyFrom(buffer, pos, size);
-        pos += size;
-        return result;
-      }
-      if (size == 0) {
-        return ByteString.EMPTY;
-      }
-      // Slow path:  Build a byte array first then copy it.
-      return ByteString.wrap(readRawBytesSlowPath(size));
-    }
-
-    @Override
-    public byte[] readByteArray() throws IOException {
-      final int size = readRawVarint32();
-      if (size <= (bufferSize - pos) && size > 0) {
-        // Fast path: We already have the bytes in a contiguous buffer, so
-        // just copy directly from it.
-        final byte[] result = Arrays.copyOfRange(buffer, pos, pos + size);
-        pos += size;
-        return result;
-      } else {
-        // Slow path: Build a byte array first then copy it.
-        return readRawBytesSlowPath(size);
-      }
-    }
-
-    @Override
-    public ByteBuffer readByteBuffer() throws IOException {
-      final int size = readRawVarint32();
-      if (size <= (bufferSize - pos) && size > 0) {
-        // Fast path: We already have the bytes in a contiguous buffer.
-        ByteBuffer result = ByteBuffer.wrap(Arrays.copyOfRange(buffer, pos, pos + size));
-        pos += size;
-        return result;
-      }
-      if (size == 0) {
-        return Internal.EMPTY_BYTE_BUFFER;
-      }
-      // Slow path: Build a byte array first then copy it.
-      return ByteBuffer.wrap(readRawBytesSlowPath(size));
-    }
-
-    @Override
-    public int readUInt32() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public int readEnum() throws IOException {
-      return readRawVarint32();
-    }
-
-    @Override
-    public int readSFixed32() throws IOException {
-      return readRawLittleEndian32();
-    }
-
-    @Override
-    public long readSFixed64() throws IOException {
-      return readRawLittleEndian64();
-    }
-
-    @Override
-    public int readSInt32() throws IOException {
-      return decodeZigZag32(readRawVarint32());
-    }
-
-    @Override
-    public long readSInt64() throws IOException {
-      return decodeZigZag64(readRawVarint64());
-    }
-
-    // =================================================================
-
-    @Override
-    public int readRawVarint32() throws IOException {
-      // See implementation notes for readRawVarint64
-      fastpath:
-      {
-        int tempPos = pos;
-
-        if (bufferSize == tempPos) {
-          break fastpath;
-        }
-
-        final byte[] buffer = this.buffer;
-        int x;
-        if ((x = buffer[tempPos++]) >= 0) {
-          pos = tempPos;
-          return x;
-        } else if (bufferSize - tempPos < 9) {
-          break fastpath;
-        } else if ((x ^= (buffer[tempPos++] << 7)) < 0) {
-          x ^= (~0 << 7);
-        } else if ((x ^= (buffer[tempPos++] << 14)) >= 0) {
-          x ^= (~0 << 7) ^ (~0 << 14);
-        } else if ((x ^= (buffer[tempPos++] << 21)) < 0) {
-          x ^= (~0 << 7) ^ (~0 << 14) ^ (~0 << 21);
-        } else {
-          int y = buffer[tempPos++];
-          x ^= y << 28;
-          x ^= (~0 << 7) ^ (~0 << 14) ^ (~0 << 21) ^ (~0 << 28);
-          if (y < 0
-              && buffer[tempPos++] < 0
-              && buffer[tempPos++] < 0
-              && buffer[tempPos++] < 0
-              && buffer[tempPos++] < 0
-              && buffer[tempPos++] < 0) {
-            break fastpath; // Will throw malformedVarint()
-          }
-        }
-        pos = tempPos;
-        return x;
-      }
-      return (int) readRawVarint64SlowPath();
-    }
-
-    private void skipRawVarint() throws IOException {
-      if (bufferSize - pos >= MAX_VARINT_SIZE) {
-        skipRawVarintFastPath();
-      } else {
-        skipRawVarintSlowPath();
-      }
-    }
-
-    private void skipRawVarintFastPath() throws IOException {
-      for (int i = 0; i < MAX_VARINT_SIZE; i++) {
-        if (buffer[pos++] >= 0) {
-          return;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    private void skipRawVarintSlowPath() throws IOException {
-      for (int i = 0; i < MAX_VARINT_SIZE; i++) {
-        if (readRawByte() >= 0) {
-          return;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    @Override
-    public long readRawVarint64() throws IOException {
-      // Implementation notes:
-      //
-      // Optimized for one-byte values, expected to be common.
-      // The particular code below was selected from various candidates
-      // empirically, by winning VarintBenchmark.
-      //
-      // Sign extension of (signed) Java bytes is usually a nuisance, but
-      // we exploit it here to more easily obtain the sign of bytes read.
-      // Instead of cleaning up the sign extension bits by masking eagerly,
-      // we delay until we find the final (positive) byte, when we clear all
-      // accumulated bits with one xor.  We depend on javac to constant fold.
-      fastpath:
-      {
-        int tempPos = pos;
-
-        if (bufferSize == tempPos) {
-          break fastpath;
-        }
-
-        final byte[] buffer = this.buffer;
-        long x;
-        int y;
-        if ((y = buffer[tempPos++]) >= 0) {
-          pos = tempPos;
-          return y;
-        } else if (bufferSize - tempPos < 9) {
-          break fastpath;
-        } else if ((y ^= (buffer[tempPos++] << 7)) < 0) {
-          x = y ^ (~0 << 7);
-        } else if ((y ^= (buffer[tempPos++] << 14)) >= 0) {
-          x = y ^ ((~0 << 7) ^ (~0 << 14));
-        } else if ((y ^= (buffer[tempPos++] << 21)) < 0) {
-          x = y ^ ((~0 << 7) ^ (~0 << 14) ^ (~0 << 21));
-        } else if ((x = y ^ ((long) buffer[tempPos++] << 28)) >= 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28);
-        } else if ((x ^= ((long) buffer[tempPos++] << 35)) < 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28) ^ (~0L << 35);
-        } else if ((x ^= ((long) buffer[tempPos++] << 42)) >= 0L) {
-          x ^= (~0L << 7) ^ (~0L << 14) ^ (~0L << 21) ^ (~0L << 28) ^ (~0L << 35) ^ (~0L << 42);
-        } else if ((x ^= ((long) buffer[tempPos++] << 49)) < 0L) {
-          x ^=
-              (~0L << 7)
-                  ^ (~0L << 14)
-                  ^ (~0L << 21)
-                  ^ (~0L << 28)
-                  ^ (~0L << 35)
-                  ^ (~0L << 42)
-                  ^ (~0L << 49);
-        } else {
-          x ^= ((long) buffer[tempPos++] << 56);
-          x ^=
-              (~0L << 7)
-                  ^ (~0L << 14)
-                  ^ (~0L << 21)
-                  ^ (~0L << 28)
-                  ^ (~0L << 35)
-                  ^ (~0L << 42)
-                  ^ (~0L << 49)
-                  ^ (~0L << 56);
-          if (x < 0L) {
-            if (buffer[tempPos++] < 0L) {
-              break fastpath; // Will throw malformedVarint()
-            }
-          }
-        }
-        pos = tempPos;
-        return x;
-      }
-      return readRawVarint64SlowPath();
-    }
-
-    @Override
-    long readRawVarint64SlowPath() throws IOException {
-      long result = 0;
-      for (int shift = 0; shift < 64; shift += 7) {
-        final byte b = readRawByte();
-        result |= (long) (b & 0x7F) << shift;
-        if ((b & 0x80) == 0) {
-          return result;
-        }
-      }
-      throw InvalidProtocolBufferException.malformedVarint();
-    }
-
-    @Override
-    public int readRawLittleEndian32() throws IOException {
-      int tempPos = pos;
-
-      if (bufferSize - tempPos < FIXED_32_SIZE) {
-        refillBuffer(FIXED_32_SIZE);
-        tempPos = pos;
-      }
-
-      final byte[] buffer = this.buffer;
-      pos = tempPos + FIXED_32_SIZE;
-      return (((buffer[tempPos] & 0xff))
-          | ((buffer[tempPos + 1] & 0xff) << 8)
-          | ((buffer[tempPos + 2] & 0xff) << 16)
-          | ((buffer[tempPos + 3] & 0xff) << 24));
-    }
-
-    @Override
-    public long readRawLittleEndian64() throws IOException {
-      int tempPos = pos;
-
-      if (bufferSize - tempPos < FIXED_64_SIZE) {
-        refillBuffer(FIXED_64_SIZE);
-        tempPos = pos;
-      }
-
-      final byte[] buffer = this.buffer;
-      pos = tempPos + FIXED_64_SIZE;
-      return (((buffer[tempPos] & 0xffL))
-          | ((buffer[tempPos + 1] & 0xffL) << 8)
-          | ((buffer[tempPos + 2] & 0xffL) << 16)
-          | ((buffer[tempPos + 3] & 0xffL) << 24)
-          | ((buffer[tempPos + 4] & 0xffL) << 32)
-          | ((buffer[tempPos + 5] & 0xffL) << 40)
-          | ((buffer[tempPos + 6] & 0xffL) << 48)
-          | ((buffer[tempPos + 7] & 0xffL) << 56));
-    }
-
-    // -----------------------------------------------------------------
-
-    @Override
-    public void enableAliasing(boolean enabled) {
-      // TODO(nathanmittler): Ideally we should throw here. Do nothing for backward compatibility.
-    }
-
-    @Override
-    public void resetSizeCounter() {
-      totalBytesRetired = -pos;
-    }
-
-    @Override
-    public int pushLimit(int byteLimit) throws InvalidProtocolBufferException {
-      if (byteLimit < 0) {
-        throw InvalidProtocolBufferException.negativeSize();
-      }
-      byteLimit += totalBytesRetired + pos;
-      final int oldLimit = currentLimit;
-      if (byteLimit > oldLimit) {
-        throw InvalidProtocolBufferException.truncatedMessage();
-      }
-      currentL

<TRUNCATED>

[02/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
deleted file mode 100644
index 3d2ad2e..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/AdminProtos.java
+++ /dev/null
@@ -1,28175 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: Admin.proto
-
-package org.apache.hadoop.hbase.shaded.protobuf.generated;
-
-public final class AdminProtos {
-  private AdminProtos() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  public interface GetRegionInfoRequestOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:hbase.pb.GetRegionInfoRequest)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    boolean hasRegion();
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
-
-    /**
-     * <code>optional bool compaction_state = 2;</code>
-     */
-    boolean hasCompactionState();
-    /**
-     * <code>optional bool compaction_state = 2;</code>
-     */
-    boolean getCompactionState();
-
-    /**
-     * <code>optional bool best_split_row = 3;</code>
-     */
-    boolean hasBestSplitRow();
-    /**
-     * <code>optional bool best_split_row = 3;</code>
-     */
-    boolean getBestSplitRow();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.GetRegionInfoRequest}
-   */
-  public  static final class GetRegionInfoRequest extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:hbase.pb.GetRegionInfoRequest)
-      GetRegionInfoRequestOrBuilder {
-    // Use GetRegionInfoRequest.newBuilder() to construct.
-    private GetRegionInfoRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private GetRegionInfoRequest() {
-      compactionState_ = false;
-      bestSplitRow_ = false;
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private GetRegionInfoRequest(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = region_.toBuilder();
-              }
-              region_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(region_);
-                region_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000001;
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              compactionState_ = input.readBool();
-              break;
-            }
-            case 24: {
-              bitField0_ |= 0x00000004;
-              bestSplitRow_ = input.readBool();
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int REGION_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_;
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    public boolean hasRegion() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
-      return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
-    }
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
-      return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
-    }
-
-    public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
-    private boolean compactionState_;
-    /**
-     * <code>optional bool compaction_state = 2;</code>
-     */
-    public boolean hasCompactionState() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional bool compaction_state = 2;</code>
-     */
-    public boolean getCompactionState() {
-      return compactionState_;
-    }
-
-    public static final int BEST_SPLIT_ROW_FIELD_NUMBER = 3;
-    private boolean bestSplitRow_;
-    /**
-     * <code>optional bool best_split_row = 3;</code>
-     */
-    public boolean hasBestSplitRow() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>optional bool best_split_row = 3;</code>
-     */
-    public boolean getBestSplitRow() {
-      return bestSplitRow_;
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (!hasRegion()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!getRegion().isInitialized()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, getRegion());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeBool(2, compactionState_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeBool(3, bestSplitRow_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, getRegion());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeBoolSize(2, compactionState_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeBoolSize(3, bestSplitRow_);
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest) obj;
-
-      boolean result = true;
-      result = result && (hasRegion() == other.hasRegion());
-      if (hasRegion()) {
-        result = result && getRegion()
-            .equals(other.getRegion());
-      }
-      result = result && (hasCompactionState() == other.hasCompactionState());
-      if (hasCompactionState()) {
-        result = result && (getCompactionState()
-            == other.getCompactionState());
-      }
-      result = result && (hasBestSplitRow() == other.hasBestSplitRow());
-      if (hasBestSplitRow()) {
-        result = result && (getBestSplitRow()
-            == other.getBestSplitRow());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasRegion()) {
-        hash = (37 * hash) + REGION_FIELD_NUMBER;
-        hash = (53 * hash) + getRegion().hashCode();
-      }
-      if (hasCompactionState()) {
-        hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
-        hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-            getCompactionState());
-      }
-      if (hasBestSplitRow()) {
-        hash = (37 * hash) + BEST_SPLIT_ROW_FIELD_NUMBER;
-        hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-            getBestSplitRow());
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.GetRegionInfoRequest}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:hbase.pb.GetRegionInfoRequest)
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequestOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-          getRegionFieldBuilder();
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        if (regionBuilder_ == null) {
-          region_ = null;
-        } else {
-          regionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        compactionState_ = false;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        bestSplitRow_ = false;
-        bitField0_ = (bitField0_ & ~0x00000004);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest build() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest buildPartial() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        if (regionBuilder_ == null) {
-          result.region_ = region_;
-        } else {
-          result.region_ = regionBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.compactionState_ = compactionState_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        result.bestSplitRow_ = bestSplitRow_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest other) {
-        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest.getDefaultInstance()) return this;
-        if (other.hasRegion()) {
-          mergeRegion(other.getRegion());
-        }
-        if (other.hasCompactionState()) {
-          setCompactionState(other.getCompactionState());
-        }
-        if (other.hasBestSplitRow()) {
-          setBestSplitRow(other.getBestSplitRow());
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasRegion()) {
-          return false;
-        }
-        if (!getRegion().isInitialized()) {
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_ = null;
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_;
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      public boolean hasRegion() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
-        if (regionBuilder_ == null) {
-          return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
-        } else {
-          return regionBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      public Builder setRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
-        if (regionBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          region_ = value;
-          onChanged();
-        } else {
-          regionBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      public Builder setRegion(
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
-        if (regionBuilder_ == null) {
-          region_ = builderForValue.build();
-          onChanged();
-        } else {
-          regionBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      public Builder mergeRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
-        if (regionBuilder_ == null) {
-          if (((bitField0_ & 0x00000001) == 0x00000001) &&
-              region_ != null &&
-              region_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
-            region_ =
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial();
-          } else {
-            region_ = value;
-          }
-          onChanged();
-        } else {
-          regionBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      public Builder clearRegion() {
-        if (regionBuilder_ == null) {
-          region_ = null;
-          onChanged();
-        } else {
-          regionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() {
-        bitField0_ |= 0x00000001;
-        onChanged();
-        return getRegionFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
-        if (regionBuilder_ != null) {
-          return regionBuilder_.getMessageOrBuilder();
-        } else {
-          return region_ == null ?
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
-        }
-      }
-      /**
-       * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-       */
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> 
-          getRegionFieldBuilder() {
-        if (regionBuilder_ == null) {
-          regionBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
-                  getRegion(),
-                  getParentForChildren(),
-                  isClean());
-          region_ = null;
-        }
-        return regionBuilder_;
-      }
-
-      private boolean compactionState_ ;
-      /**
-       * <code>optional bool compaction_state = 2;</code>
-       */
-      public boolean hasCompactionState() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional bool compaction_state = 2;</code>
-       */
-      public boolean getCompactionState() {
-        return compactionState_;
-      }
-      /**
-       * <code>optional bool compaction_state = 2;</code>
-       */
-      public Builder setCompactionState(boolean value) {
-        bitField0_ |= 0x00000002;
-        compactionState_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bool compaction_state = 2;</code>
-       */
-      public Builder clearCompactionState() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        compactionState_ = false;
-        onChanged();
-        return this;
-      }
-
-      private boolean bestSplitRow_ ;
-      /**
-       * <code>optional bool best_split_row = 3;</code>
-       */
-      public boolean hasBestSplitRow() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
-      }
-      /**
-       * <code>optional bool best_split_row = 3;</code>
-       */
-      public boolean getBestSplitRow() {
-        return bestSplitRow_;
-      }
-      /**
-       * <code>optional bool best_split_row = 3;</code>
-       */
-      public Builder setBestSplitRow(boolean value) {
-        bitField0_ |= 0x00000004;
-        bestSplitRow_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bool best_split_row = 3;</code>
-       */
-      public Builder clearBestSplitRow() {
-        bitField0_ = (bitField0_ & ~0x00000004);
-        bestSplitRow_ = false;
-        onChanged();
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.GetRegionInfoRequest)
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.GetRegionInfoRequest)
-    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoRequest>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<GetRegionInfoRequest>() {
-      public GetRegionInfoRequest parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new GetRegionInfoRequest(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoRequest> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoRequest> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoRequest getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface GetRegionInfoResponseOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:hbase.pb.GetRegionInfoResponse)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-     */
-    boolean hasRegionInfo();
-    /**
-     * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo();
-    /**
-     * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder();
-
-    /**
-     * <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
-     */
-    boolean hasCompactionState();
-    /**
-     * <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState();
-
-    /**
-     * <code>optional bool isRecovering = 3;</code>
-     */
-    boolean hasIsRecovering();
-    /**
-     * <code>optional bool isRecovering = 3;</code>
-     */
-    boolean getIsRecovering();
-
-    /**
-     * <pre>
-     * True if region is splittable, false otherwise.
-     * </pre>
-     *
-     * <code>optional bool splittable = 4;</code>
-     */
-    boolean hasSplittable();
-    /**
-     * <pre>
-     * True if region is splittable, false otherwise.
-     * </pre>
-     *
-     * <code>optional bool splittable = 4;</code>
-     */
-    boolean getSplittable();
-
-    /**
-     * <pre>
-     * True if region is mergeable, false otherwise.
-     * </pre>
-     *
-     * <code>optional bool mergeable = 5;</code>
-     */
-    boolean hasMergeable();
-    /**
-     * <pre>
-     * True if region is mergeable, false otherwise.
-     * </pre>
-     *
-     * <code>optional bool mergeable = 5;</code>
-     */
-    boolean getMergeable();
-
-    /**
-     * <pre>
-     * Get bestSplitRow
-     * </pre>
-     *
-     * <code>optional bytes best_split_row = 6;</code>
-     */
-    boolean hasBestSplitRow();
-    /**
-     * <pre>
-     * Get bestSplitRow
-     * </pre>
-     *
-     * <code>optional bytes best_split_row = 6;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getBestSplitRow();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.GetRegionInfoResponse}
-   */
-  public  static final class GetRegionInfoResponse extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:hbase.pb.GetRegionInfoResponse)
-      GetRegionInfoResponseOrBuilder {
-    // Use GetRegionInfoResponse.newBuilder() to construct.
-    private GetRegionInfoResponse(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private GetRegionInfoResponse() {
-      compactionState_ = 0;
-      isRecovering_ = false;
-      splittable_ = false;
-      mergeable_ = false;
-      bestSplitRow_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private GetRegionInfoResponse(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = regionInfo_.toBuilder();
-              }
-              regionInfo_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(regionInfo_);
-                regionInfo_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000001;
-              break;
-            }
-            case 16: {
-              int rawValue = input.readEnum();
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState value = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.valueOf(rawValue);
-              if (value == null) {
-                unknownFields.mergeVarintField(2, rawValue);
-              } else {
-                bitField0_ |= 0x00000002;
-                compactionState_ = rawValue;
-              }
-              break;
-            }
-            case 24: {
-              bitField0_ |= 0x00000004;
-              isRecovering_ = input.readBool();
-              break;
-            }
-            case 32: {
-              bitField0_ |= 0x00000008;
-              splittable_ = input.readBool();
-              break;
-            }
-            case 40: {
-              bitField0_ |= 0x00000010;
-              mergeable_ = input.readBool();
-              break;
-            }
-            case 50: {
-              bitField0_ |= 0x00000020;
-              bestSplitRow_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.Builder.class);
-    }
-
-    /**
-     * Protobuf enum {@code hbase.pb.GetRegionInfoResponse.CompactionState}
-     */
-    public enum CompactionState
-        implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
-      /**
-       * <code>NONE = 0;</code>
-       */
-      NONE(0),
-      /**
-       * <code>MINOR = 1;</code>
-       */
-      MINOR(1),
-      /**
-       * <code>MAJOR = 2;</code>
-       */
-      MAJOR(2),
-      /**
-       * <code>MAJOR_AND_MINOR = 3;</code>
-       */
-      MAJOR_AND_MINOR(3),
-      ;
-
-      /**
-       * <code>NONE = 0;</code>
-       */
-      public static final int NONE_VALUE = 0;
-      /**
-       * <code>MINOR = 1;</code>
-       */
-      public static final int MINOR_VALUE = 1;
-      /**
-       * <code>MAJOR = 2;</code>
-       */
-      public static final int MAJOR_VALUE = 2;
-      /**
-       * <code>MAJOR_AND_MINOR = 3;</code>
-       */
-      public static final int MAJOR_AND_MINOR_VALUE = 3;
-
-
-      public final int getNumber() {
-        return value;
-      }
-
-      /**
-       * @deprecated Use {@link #forNumber(int)} instead.
-       */
-      @java.lang.Deprecated
-      public static CompactionState valueOf(int value) {
-        return forNumber(value);
-      }
-
-      public static CompactionState forNumber(int value) {
-        switch (value) {
-          case 0: return NONE;
-          case 1: return MINOR;
-          case 2: return MAJOR;
-          case 3: return MAJOR_AND_MINOR;
-          default: return null;
-        }
-      }
-
-      public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CompactionState>
-          internalGetValueMap() {
-        return internalValueMap;
-      }
-      private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
-          CompactionState> internalValueMap =
-            new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<CompactionState>() {
-              public CompactionState findValueByNumber(int number) {
-                return CompactionState.forNumber(number);
-              }
-            };
-
-      public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
-          getValueDescriptor() {
-        return getDescriptor().getValues().get(ordinal());
-      }
-      public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-          getDescriptorForType() {
-        return getDescriptor();
-      }
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.getDescriptor().getEnumTypes().get(0);
-      }
-
-      private static final CompactionState[] VALUES = values();
-
-      public static CompactionState valueOf(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-        if (desc.getType() != getDescriptor()) {
-          throw new java.lang.IllegalArgumentException(
-            "EnumValueDescriptor is not for this type.");
-        }
-        return VALUES[desc.getIndex()];
-      }
-
-      private final int value;
-
-      private CompactionState(int value) {
-        this.value = value;
-      }
-
-      // @@protoc_insertion_point(enum_scope:hbase.pb.GetRegionInfoResponse.CompactionState)
-    }
-
-    private int bitField0_;
-    public static final int REGION_INFO_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_;
-    /**
-     * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-     */
-    public boolean hasRegionInfo() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo() {
-      return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_;
-    }
-    /**
-     * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder() {
-      return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_;
-    }
-
-    public static final int COMPACTION_STATE_FIELD_NUMBER = 2;
-    private int compactionState_;
-    /**
-     * <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
-     */
-    public boolean hasCompactionState() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState() {
-      org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState result = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.valueOf(compactionState_);
-      return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE : result;
-    }
-
-    public static final int ISRECOVERING_FIELD_NUMBER = 3;
-    private boolean isRecovering_;
-    /**
-     * <code>optional bool isRecovering = 3;</code>
-     */
-    public boolean hasIsRecovering() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>optional bool isRecovering = 3;</code>
-     */
-    public boolean getIsRecovering() {
-      return isRecovering_;
-    }
-
-    public static final int SPLITTABLE_FIELD_NUMBER = 4;
-    private boolean splittable_;
-    /**
-     * <pre>
-     * True if region is splittable, false otherwise.
-     * </pre>
-     *
-     * <code>optional bool splittable = 4;</code>
-     */
-    public boolean hasSplittable() {
-      return ((bitField0_ & 0x00000008) == 0x00000008);
-    }
-    /**
-     * <pre>
-     * True if region is splittable, false otherwise.
-     * </pre>
-     *
-     * <code>optional bool splittable = 4;</code>
-     */
-    public boolean getSplittable() {
-      return splittable_;
-    }
-
-    public static final int MERGEABLE_FIELD_NUMBER = 5;
-    private boolean mergeable_;
-    /**
-     * <pre>
-     * True if region is mergeable, false otherwise.
-     * </pre>
-     *
-     * <code>optional bool mergeable = 5;</code>
-     */
-    public boolean hasMergeable() {
-      return ((bitField0_ & 0x00000010) == 0x00000010);
-    }
-    /**
-     * <pre>
-     * True if region is mergeable, false otherwise.
-     * </pre>
-     *
-     * <code>optional bool mergeable = 5;</code>
-     */
-    public boolean getMergeable() {
-      return mergeable_;
-    }
-
-    public static final int BEST_SPLIT_ROW_FIELD_NUMBER = 6;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bestSplitRow_;
-    /**
-     * <pre>
-     * Get bestSplitRow
-     * </pre>
-     *
-     * <code>optional bytes best_split_row = 6;</code>
-     */
-    public boolean hasBestSplitRow() {
-      return ((bitField0_ & 0x00000020) == 0x00000020);
-    }
-    /**
-     * <pre>
-     * Get bestSplitRow
-     * </pre>
-     *
-     * <code>optional bytes best_split_row = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getBestSplitRow() {
-      return bestSplitRow_;
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (!hasRegionInfo()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!getRegionInfo().isInitialized()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, getRegionInfo());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeEnum(2, compactionState_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeBool(3, isRecovering_);
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        output.writeBool(4, splittable_);
-      }
-      if (((bitField0_ & 0x00000010) == 0x00000010)) {
-        output.writeBool(5, mergeable_);
-      }
-      if (((bitField0_ & 0x00000020) == 0x00000020)) {
-        output.writeBytes(6, bestSplitRow_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, getRegionInfo());
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeEnumSize(2, compactionState_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeBoolSize(3, isRecovering_);
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeBoolSize(4, splittable_);
-      }
-      if (((bitField0_ & 0x00000010) == 0x00000010)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeBoolSize(5, mergeable_);
-      }
-      if (((bitField0_ & 0x00000020) == 0x00000020)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeBytesSize(6, bestSplitRow_);
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse) obj;
-
-      boolean result = true;
-      result = result && (hasRegionInfo() == other.hasRegionInfo());
-      if (hasRegionInfo()) {
-        result = result && getRegionInfo()
-            .equals(other.getRegionInfo());
-      }
-      result = result && (hasCompactionState() == other.hasCompactionState());
-      if (hasCompactionState()) {
-        result = result && compactionState_ == other.compactionState_;
-      }
-      result = result && (hasIsRecovering() == other.hasIsRecovering());
-      if (hasIsRecovering()) {
-        result = result && (getIsRecovering()
-            == other.getIsRecovering());
-      }
-      result = result && (hasSplittable() == other.hasSplittable());
-      if (hasSplittable()) {
-        result = result && (getSplittable()
-            == other.getSplittable());
-      }
-      result = result && (hasMergeable() == other.hasMergeable());
-      if (hasMergeable()) {
-        result = result && (getMergeable()
-            == other.getMergeable());
-      }
-      result = result && (hasBestSplitRow() == other.hasBestSplitRow());
-      if (hasBestSplitRow()) {
-        result = result && getBestSplitRow()
-            .equals(other.getBestSplitRow());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasRegionInfo()) {
-        hash = (37 * hash) + REGION_INFO_FIELD_NUMBER;
-        hash = (53 * hash) + getRegionInfo().hashCode();
-      }
-      if (hasCompactionState()) {
-        hash = (37 * hash) + COMPACTION_STATE_FIELD_NUMBER;
-        hash = (53 * hash) + compactionState_;
-      }
-      if (hasIsRecovering()) {
-        hash = (37 * hash) + ISRECOVERING_FIELD_NUMBER;
-        hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-            getIsRecovering());
-      }
-      if (hasSplittable()) {
-        hash = (37 * hash) + SPLITTABLE_FIELD_NUMBER;
-        hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-            getSplittable());
-      }
-      if (hasMergeable()) {
-        hash = (37 * hash) + MERGEABLE_FIELD_NUMBER;
-        hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-            getMergeable());
-      }
-      if (hasBestSplitRow()) {
-        hash = (37 * hash) + BEST_SPLIT_ROW_FIELD_NUMBER;
-        hash = (53 * hash) + getBestSplitRow().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.GetRegionInfoResponse}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:hbase.pb.GetRegionInfoResponse)
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponseOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-          getRegionInfoFieldBuilder();
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        if (regionInfoBuilder_ == null) {
-          regionInfo_ = null;
-        } else {
-          regionInfoBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        compactionState_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        isRecovering_ = false;
-        bitField0_ = (bitField0_ & ~0x00000004);
-        splittable_ = false;
-        bitField0_ = (bitField0_ & ~0x00000008);
-        mergeable_ = false;
-        bitField0_ = (bitField0_ & ~0x00000010);
-        bestSplitRow_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000020);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetRegionInfoResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse build() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse buildPartial() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        if (regionInfoBuilder_ == null) {
-          result.regionInfo_ = regionInfo_;
-        } else {
-          result.regionInfo_ = regionInfoBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.compactionState_ = compactionState_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        result.isRecovering_ = isRecovering_;
-        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
-          to_bitField0_ |= 0x00000008;
-        }
-        result.splittable_ = splittable_;
-        if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
-          to_bitField0_ |= 0x00000010;
-        }
-        result.mergeable_ = mergeable_;
-        if (((from_bitField0_ & 0x00000020) == 0x00000020)) {
-          to_bitField0_ |= 0x00000020;
-        }
-        result.bestSplitRow_ = bestSplitRow_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse other) {
-        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.getDefaultInstance()) return this;
-        if (other.hasRegionInfo()) {
-          mergeRegionInfo(other.getRegionInfo());
-        }
-        if (other.hasCompactionState()) {
-          setCompactionState(other.getCompactionState());
-        }
-        if (other.hasIsRecovering()) {
-          setIsRecovering(other.getIsRecovering());
-        }
-        if (other.hasSplittable()) {
-          setSplittable(other.getSplittable());
-        }
-        if (other.hasMergeable()) {
-          setMergeable(other.getMergeable());
-        }
-        if (other.hasBestSplitRow()) {
-          setBestSplitRow(other.getBestSplitRow());
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasRegionInfo()) {
-          return false;
-        }
-        if (!getRegionInfo().isInitialized()) {
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_ = null;
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_;
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      public boolean hasRegionInfo() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo() {
-        if (regionInfoBuilder_ == null) {
-          return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_;
-        } else {
-          return regionInfoBuilder_.getMessage();
-        }
-      }
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      public Builder setRegionInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo value) {
-        if (regionInfoBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          regionInfo_ = value;
-          onChanged();
-        } else {
-          regionInfoBuilder_.setMessage(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      public Builder setRegionInfo(
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder builderForValue) {
-        if (regionInfoBuilder_ == null) {
-          regionInfo_ = builderForValue.build();
-          onChanged();
-        } else {
-          regionInfoBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      public Builder mergeRegionInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo value) {
-        if (regionInfoBuilder_ == null) {
-          if (((bitField0_ & 0x00000001) == 0x00000001) &&
-              regionInfo_ != null &&
-              regionInfo_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance()) {
-            regionInfo_ =
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.newBuilder(regionInfo_).mergeFrom(value).buildPartial();
-          } else {
-            regionInfo_ = value;
-          }
-          onChanged();
-        } else {
-          regionInfoBuilder_.mergeFrom(value);
-        }
-        bitField0_ |= 0x00000001;
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      public Builder clearRegionInfo() {
-        if (regionInfoBuilder_ == null) {
-          regionInfo_ = null;
-          onChanged();
-        } else {
-          regionInfoBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder getRegionInfoBuilder() {
-        bitField0_ |= 0x00000001;
-        onChanged();
-        return getRegionInfoFieldBuilder().getBuilder();
-      }
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder() {
-        if (regionInfoBuilder_ != null) {
-          return regionInfoBuilder_.getMessageOrBuilder();
-        } else {
-          return regionInfo_ == null ?
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_;
-        }
-      }
-      /**
-       * <code>required .hbase.pb.RegionInfo region_info = 1;</code>
-       */
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> 
-          getRegionInfoFieldBuilder() {
-        if (regionInfoBuilder_ == null) {
-          regionInfoBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>(
-                  getRegionInfo(),
-                  getParentForChildren(),
-                  isClean());
-          regionInfo_ = null;
-        }
-        return regionInfoBuilder_;
-      }
-
-      private int compactionState_ = 0;
-      /**
-       * <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
-       */
-      public boolean hasCompactionState() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState getCompactionState() {
-        org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState result = org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.valueOf(compactionState_);
-        return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState.NONE : result;
-      }
-      /**
-       * <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
-       */
-      public Builder setCompactionState(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState value) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        bitField0_ |= 0x00000002;
-        compactionState_ = value.getNumber();
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional .hbase.pb.GetRegionInfoResponse.CompactionState compaction_state = 2;</code>
-       */
-      public Builder clearCompactionState() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        compactionState_ = 0;
-        onChanged();
-        return this;
-      }
-
-      private boolean isRecovering_ ;
-      /**
-       * <code>optional bool isRecovering = 3;</code>
-       */
-      public boolean hasIsRecovering() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
-      }
-      /**
-       * <code>optional bool isRecovering = 3;</code>
-       */
-      public boolean getIsRecovering() {
-        return isRecovering_;
-      }
-      /**
-       * <code>optional bool isRecovering = 3;</code>
-       */
-      public Builder setIsRecovering(boolean value) {
-        bitField0_ |= 0x00000004;
-        isRecovering_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bool isRecovering = 3;</code>
-       */
-      public Builder clearIsRecovering() {
-        bitField0_ = (bitField0_ & ~0x00000004);
-        isRecovering_ = false;
-        onChanged();
-        return this;
-      }
-
-      private boolean splittable_ ;
-      /**
-       * <pre>
-       * True if region is splittable, false otherwise.
-       * </pre>
-       *
-       * <code>optional bool splittable = 4;</code>
-       */
-      public boolean hasSplittable() {
-        return ((bitField0_ & 0x00000008) == 0x00000008);
-      }
-      /**
-       * <pre>
-       * True if region is splittable, false otherwise.
-       * </pre>
-       *
-       * <code>optional bool splittable = 4;</code>
-       */
-      public boolean getSplittable() {
-        return splittable_;
-      }
-      /**
-       * <pre>
-       * True if region is splittable, false otherwise.
-       * </pre>
-       *
-       * <code>optional bool splittable = 4;</code>
-       */
-      public Builder setSplittable(boolean value) {
-        bitField0_ |= 0x00000008;
-        splittable_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * True if region is splittable, false otherwise.
-       * </pre>
-       *
-       * <code>optional bool splittable = 4;</code>
-       */
-      public Builder clearSplittable() {
-        bitField0_ = (bitField0_ & ~0x00000008);
-        splittable_ = false;
-        onChanged();
-        return this;
-      }
-
-      private boolean mergeable_ ;
-      /**
-       * <pre>
-       * True if region is mergeable, false otherwise.
-       * </pre>
-       *
-       * <code>optional bool mergeable = 5;</code>
-       */
-      public boolean hasMergeable() {
-        return ((bitField0_ & 0x00000010) == 0x00000010);
-      }
-      /**
-       * <pre>
-       * True if region is mergeable, false otherwise.
-       * </pre>
-       *
-       * <code>optional bool mergeable = 5;</code>
-       */
-      public boolean getMergeable() {
-        return mergeable_;
-      }
-      /**
-       * <pre>
-       * True if region is mergeable, false otherwise.
-       * </pre>
-       *
-       * <code>optional bool mergeable = 5;</code>
-       */
-      public Builder setMergeable(boolean value) {
-        bitField0_ |= 0x00000010;
-        mergeable_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * True if region is mergeable, false otherwise.
-       * </pre>
-       *
-       * <code>optional bool mergeable = 5;</code>
-       */
-      public Builder clearMergeable() {
-        bitField0_ = (bitField0_ & ~0x00000010);
-        mergeable_ = false;
-        onChanged();
-        return this;
-      }
-
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bestSplitRow_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <pre>
-       * Get bestSplitRow
-       * </pre>
-       *
-       * <code>optional bytes best_split_row = 6;</code>
-       */
-      public boolean hasBestSplitRow() {
-        return ((bitField0_ & 0x00000020) == 0x00000020);
-      }
-      /**
-       * <pre>
-       * Get bestSplitRow
-       * </pre>
-       *
-       * <code>optional bytes best_split_row = 6;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getBestSplitRow() {
-        return bestSplitRow_;
-      }
-      /**
-       * <pre>
-       * Get bestSplitRow
-       * </pre>
-       *
-       * <code>optional bytes best_split_row = 6;</code>
-       */
-      public Builder setBestSplitRow(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000020;
-        bestSplitRow_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * Get bestSplitRow
-       * </pre>
-       *
-       * <code>optional bytes best_split_row = 6;</code>
-       */
-      public Builder clearBestSplitRow() {
-        bitField0_ = (bitField0_ & ~0x00000020);
-        bestSplitRow_ = getDefaultInstance().getBestSplitRow();
-        onChanged();
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.GetRegionInfoResponse)
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.GetRegionInfoResponse)
-    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoResponse>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<GetRegionInfoResponse>() {
-      public GetRegionInfoResponse parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new GetRegionInfoResponse(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoResponse> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<GetRegionInfoResponse> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface GetStoreFileRequestOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:hbase.pb.GetStoreFileRequest)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    boolean hasRegion();
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
-    /**
-     * <code>required .hbase.pb.RegionSpecifier region = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
-
-    /**
-     * <code>repeated bytes family = 2;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString> getFamilyList();
-    /**
-     * <code>repeated bytes family = 2;</code>
-     */
-    int getFamilyCount();
-    /**
-     * <code>repeated bytes family = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getFamily(int index);
-  }
-  /**
-   * <pre>
-   **
-   * Get a list of store files for a set of column families in a particular region.
-   * If no column family is specified, get the store files for all column families.
-   * </pre>
-   *
-   * Protobuf type {@code hbase.pb.GetStoreFileRequest}
-   */
-  public  static final class GetStoreFileRequest extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:hbase.pb.GetStoreFileRequest)
-      GetStoreFileRequestOrBuilder {
-    // Use GetStoreFileRequest.newBuilder() to construct.
-    private GetStoreFileRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private GetStoreFileRequest() {
-      family_ = java.util.Collections.emptyList();
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private GetStoreFileRequest(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = region_.toBuilder();
-              }
-              region_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(region_);
-                region_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000001;
-              break;
-            }
-            case 18: {
-              if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-                family_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString>();
-                mutable_bitField0_ |= 0x00000002;
-              }
-              family_.add(input.readBytes());
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-          family_ = java.util.Collections.unmodifiableList(family_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.internal_static_hbase_pb_GetStoreFileRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetStoreFileRequest.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int REGION_FIEL

<TRUNCATED>

[08/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Value.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Value.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Value.java
deleted file mode 100644
index 593043b..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Value.java
+++ /dev/null
@@ -1,1410 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/struct.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * `Value` represents a dynamically typed value which can be either
- * null, a number, a string, a boolean, a recursive struct value, or a
- * list of values. A producer of value is expected to set one of that
- * variants, absence of any variant indicates an error.
- * The JSON representation for `Value` is JSON value.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Value}
- */
-public  final class Value extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Value)
-    ValueOrBuilder {
-  // Use Value.newBuilder() to construct.
-  private Value(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Value() {
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Value(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-            int rawValue = input.readEnum();
-            kindCase_ = 1;
-            kind_ = rawValue;
-            break;
-          }
-          case 17: {
-            kindCase_ = 2;
-            kind_ = input.readDouble();
-            break;
-          }
-          case 26: {
-            java.lang.String s = input.readStringRequireUtf8();
-            kindCase_ = 3;
-            kind_ = s;
-            break;
-          }
-          case 32: {
-            kindCase_ = 4;
-            kind_ = input.readBool();
-            break;
-          }
-          case 42: {
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.Builder subBuilder = null;
-            if (kindCase_ == 5) {
-              subBuilder = ((org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_).toBuilder();
-            }
-            kind_ =
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.parser(), extensionRegistry);
-            if (subBuilder != null) {
-              subBuilder.mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_);
-              kind_ = subBuilder.buildPartial();
-            }
-            kindCase_ = 5;
-            break;
-          }
-          case 50: {
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.Builder subBuilder = null;
-            if (kindCase_ == 6) {
-              subBuilder = ((org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_).toBuilder();
-            }
-            kind_ =
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.parser(), extensionRegistry);
-            if (subBuilder != null) {
-              subBuilder.mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_);
-              kind_ = subBuilder.buildPartial();
-            }
-            kindCase_ = 6;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Value_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Value_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder.class);
-  }
-
-  private int kindCase_ = 0;
-  private java.lang.Object kind_;
-  public enum KindCase
-      implements org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLite {
-    NULL_VALUE(1),
-    NUMBER_VALUE(2),
-    STRING_VALUE(3),
-    BOOL_VALUE(4),
-    STRUCT_VALUE(5),
-    LIST_VALUE(6),
-    KIND_NOT_SET(0);
-    private final int value;
-    private KindCase(int value) {
-      this.value = value;
-    }
-    /**
-     * @deprecated Use {@link #forNumber(int)} instead.
-     */
-    @java.lang.Deprecated
-    public static KindCase valueOf(int value) {
-      return forNumber(value);
-    }
-
-    public static KindCase forNumber(int value) {
-      switch (value) {
-        case 1: return NULL_VALUE;
-        case 2: return NUMBER_VALUE;
-        case 3: return STRING_VALUE;
-        case 4: return BOOL_VALUE;
-        case 5: return STRUCT_VALUE;
-        case 6: return LIST_VALUE;
-        case 0: return KIND_NOT_SET;
-        default: return null;
-      }
-    }
-    public int getNumber() {
-      return this.value;
-    }
-  };
-
-  public KindCase
-  getKindCase() {
-    return KindCase.forNumber(
-        kindCase_);
-  }
-
-  public static final int NULL_VALUE_FIELD_NUMBER = 1;
-  /**
-   * <pre>
-   * Represents a null value.
-   * </pre>
-   *
-   * <code>.google.protobuf.NullValue null_value = 1;</code>
-   */
-  public int getNullValueValue() {
-    if (kindCase_ == 1) {
-      return (java.lang.Integer) kind_;
-    }
-    return 0;
-  }
-  /**
-   * <pre>
-   * Represents a null value.
-   * </pre>
-   *
-   * <code>.google.protobuf.NullValue null_value = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue getNullValue() {
-    if (kindCase_ == 1) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue result = org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue.valueOf(
-          (java.lang.Integer) kind_);
-      return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue.UNRECOGNIZED : result;
-    }
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue.NULL_VALUE;
-  }
-
-  public static final int NUMBER_VALUE_FIELD_NUMBER = 2;
-  /**
-   * <pre>
-   * Represents a double value.
-   * </pre>
-   *
-   * <code>double number_value = 2;</code>
-   */
-  public double getNumberValue() {
-    if (kindCase_ == 2) {
-      return (java.lang.Double) kind_;
-    }
-    return 0D;
-  }
-
-  public static final int STRING_VALUE_FIELD_NUMBER = 3;
-  /**
-   * <pre>
-   * Represents a string value.
-   * </pre>
-   *
-   * <code>string string_value = 3;</code>
-   */
-  public java.lang.String getStringValue() {
-    java.lang.Object ref = "";
-    if (kindCase_ == 3) {
-      ref = kind_;
-    }
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      if (kindCase_ == 3) {
-        kind_ = s;
-      }
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * Represents a string value.
-   * </pre>
-   *
-   * <code>string string_value = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getStringValueBytes() {
-    java.lang.Object ref = "";
-    if (kindCase_ == 3) {
-      ref = kind_;
-    }
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      if (kindCase_ == 3) {
-        kind_ = b;
-      }
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int BOOL_VALUE_FIELD_NUMBER = 4;
-  /**
-   * <pre>
-   * Represents a boolean value.
-   * </pre>
-   *
-   * <code>bool bool_value = 4;</code>
-   */
-  public boolean getBoolValue() {
-    if (kindCase_ == 4) {
-      return (java.lang.Boolean) kind_;
-    }
-    return false;
-  }
-
-  public static final int STRUCT_VALUE_FIELD_NUMBER = 5;
-  /**
-   * <pre>
-   * Represents a structured value.
-   * </pre>
-   *
-   * <code>.google.protobuf.Struct struct_value = 5;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct getStructValue() {
-    if (kindCase_ == 5) {
-       return (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_;
-    }
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance();
-  }
-  /**
-   * <pre>
-   * Represents a structured value.
-   * </pre>
-   *
-   * <code>.google.protobuf.Struct struct_value = 5;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.StructOrBuilder getStructValueOrBuilder() {
-    if (kindCase_ == 5) {
-       return (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_;
-    }
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance();
-  }
-
-  public static final int LIST_VALUE_FIELD_NUMBER = 6;
-  /**
-   * <pre>
-   * Represents a repeated `Value`.
-   * </pre>
-   *
-   * <code>.google.protobuf.ListValue list_value = 6;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue getListValue() {
-    if (kindCase_ == 6) {
-       return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_;
-    }
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance();
-  }
-  /**
-   * <pre>
-   * Represents a repeated `Value`.
-   * </pre>
-   *
-   * <code>.google.protobuf.ListValue list_value = 6;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValueOrBuilder getListValueOrBuilder() {
-    if (kindCase_ == 6) {
-       return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_;
-    }
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance();
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (kindCase_ == 1) {
-      output.writeEnum(1, ((java.lang.Integer) kind_));
-    }
-    if (kindCase_ == 2) {
-      output.writeDouble(
-          2, (double)((java.lang.Double) kind_));
-    }
-    if (kindCase_ == 3) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, kind_);
-    }
-    if (kindCase_ == 4) {
-      output.writeBool(
-          4, (boolean)((java.lang.Boolean) kind_));
-    }
-    if (kindCase_ == 5) {
-      output.writeMessage(5, (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_);
-    }
-    if (kindCase_ == 6) {
-      output.writeMessage(6, (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (kindCase_ == 1) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeEnumSize(1, ((java.lang.Integer) kind_));
-    }
-    if (kindCase_ == 2) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeDoubleSize(
-            2, (double)((java.lang.Double) kind_));
-    }
-    if (kindCase_ == 3) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(3, kind_);
-    }
-    if (kindCase_ == 4) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeBoolSize(
-            4, (boolean)((java.lang.Boolean) kind_));
-    }
-    if (kindCase_ == 5) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(5, (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_);
-    }
-    if (kindCase_ == 6) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(6, (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Value)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Value other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Value) obj;
-
-    boolean result = true;
-    result = result && getKindCase().equals(
-        other.getKindCase());
-    if (!result) return false;
-    switch (kindCase_) {
-      case 1:
-        result = result && getNullValueValue()
-            == other.getNullValueValue();
-        break;
-      case 2:
-        result = result && (
-            java.lang.Double.doubleToLongBits(getNumberValue())
-            == java.lang.Double.doubleToLongBits(
-                other.getNumberValue()));
-        break;
-      case 3:
-        result = result && getStringValue()
-            .equals(other.getStringValue());
-        break;
-      case 4:
-        result = result && (getBoolValue()
-            == other.getBoolValue());
-        break;
-      case 5:
-        result = result && getStructValue()
-            .equals(other.getStructValue());
-        break;
-      case 6:
-        result = result && getListValue()
-            .equals(other.getListValue());
-        break;
-      case 0:
-      default:
-    }
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    switch (kindCase_) {
-      case 1:
-        hash = (37 * hash) + NULL_VALUE_FIELD_NUMBER;
-        hash = (53 * hash) + getNullValueValue();
-        break;
-      case 2:
-        hash = (37 * hash) + NUMBER_VALUE_FIELD_NUMBER;
-        hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
-            java.lang.Double.doubleToLongBits(getNumberValue()));
-        break;
-      case 3:
-        hash = (37 * hash) + STRING_VALUE_FIELD_NUMBER;
-        hash = (53 * hash) + getStringValue().hashCode();
-        break;
-      case 4:
-        hash = (37 * hash) + BOOL_VALUE_FIELD_NUMBER;
-        hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-            getBoolValue());
-        break;
-      case 5:
-        hash = (37 * hash) + STRUCT_VALUE_FIELD_NUMBER;
-        hash = (53 * hash) + getStructValue().hashCode();
-        break;
-      case 6:
-        hash = (37 * hash) + LIST_VALUE_FIELD_NUMBER;
-        hash = (53 * hash) + getListValue().hashCode();
-        break;
-      case 0:
-      default:
-    }
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Value prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * `Value` represents a dynamically typed value which can be either
-   * null, a number, a string, a boolean, a recursive struct value, or a
-   * list of values. A producer of value is expected to set one of that
-   * variants, absence of any variant indicates an error.
-   * The JSON representation for `Value` is JSON value.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Value}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Value)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Value_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Value_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      kindCase_ = 0;
-      kind_ = null;
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Value_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Value result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Value result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Value(this);
-      if (kindCase_ == 1) {
-        result.kind_ = kind_;
-      }
-      if (kindCase_ == 2) {
-        result.kind_ = kind_;
-      }
-      if (kindCase_ == 3) {
-        result.kind_ = kind_;
-      }
-      if (kindCase_ == 4) {
-        result.kind_ = kind_;
-      }
-      if (kindCase_ == 5) {
-        if (structValueBuilder_ == null) {
-          result.kind_ = kind_;
-        } else {
-          result.kind_ = structValueBuilder_.build();
-        }
-      }
-      if (kindCase_ == 6) {
-        if (listValueBuilder_ == null) {
-          result.kind_ = kind_;
-        } else {
-          result.kind_ = listValueBuilder_.build();
-        }
-      }
-      result.kindCase_ = kindCase_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Value) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Value)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Value other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.getDefaultInstance()) return this;
-      switch (other.getKindCase()) {
-        case NULL_VALUE: {
-          setNullValueValue(other.getNullValueValue());
-          break;
-        }
-        case NUMBER_VALUE: {
-          setNumberValue(other.getNumberValue());
-          break;
-        }
-        case STRING_VALUE: {
-          kindCase_ = 3;
-          kind_ = other.kind_;
-          onChanged();
-          break;
-        }
-        case BOOL_VALUE: {
-          setBoolValue(other.getBoolValue());
-          break;
-        }
-        case STRUCT_VALUE: {
-          mergeStructValue(other.getStructValue());
-          break;
-        }
-        case LIST_VALUE: {
-          mergeListValue(other.getListValue());
-          break;
-        }
-        case KIND_NOT_SET: {
-          break;
-        }
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Value parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Value) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int kindCase_ = 0;
-    private java.lang.Object kind_;
-    public KindCase
-        getKindCase() {
-      return KindCase.forNumber(
-          kindCase_);
-    }
-
-    public Builder clearKind() {
-      kindCase_ = 0;
-      kind_ = null;
-      onChanged();
-      return this;
-    }
-
-
-    /**
-     * <pre>
-     * Represents a null value.
-     * </pre>
-     *
-     * <code>.google.protobuf.NullValue null_value = 1;</code>
-     */
-    public int getNullValueValue() {
-      if (kindCase_ == 1) {
-        return ((java.lang.Integer) kind_).intValue();
-      }
-      return 0;
-    }
-    /**
-     * <pre>
-     * Represents a null value.
-     * </pre>
-     *
-     * <code>.google.protobuf.NullValue null_value = 1;</code>
-     */
-    public Builder setNullValueValue(int value) {
-      kindCase_ = 1;
-      kind_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a null value.
-     * </pre>
-     *
-     * <code>.google.protobuf.NullValue null_value = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue getNullValue() {
-      if (kindCase_ == 1) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue result = org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue.valueOf(
-            (java.lang.Integer) kind_);
-        return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue.UNRECOGNIZED : result;
-      }
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue.NULL_VALUE;
-    }
-    /**
-     * <pre>
-     * Represents a null value.
-     * </pre>
-     *
-     * <code>.google.protobuf.NullValue null_value = 1;</code>
-     */
-    public Builder setNullValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue value) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-      kindCase_ = 1;
-      kind_ = value.getNumber();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a null value.
-     * </pre>
-     *
-     * <code>.google.protobuf.NullValue null_value = 1;</code>
-     */
-    public Builder clearNullValue() {
-      if (kindCase_ == 1) {
-        kindCase_ = 0;
-        kind_ = null;
-        onChanged();
-      }
-      return this;
-    }
-
-    /**
-     * <pre>
-     * Represents a double value.
-     * </pre>
-     *
-     * <code>double number_value = 2;</code>
-     */
-    public double getNumberValue() {
-      if (kindCase_ == 2) {
-        return (java.lang.Double) kind_;
-      }
-      return 0D;
-    }
-    /**
-     * <pre>
-     * Represents a double value.
-     * </pre>
-     *
-     * <code>double number_value = 2;</code>
-     */
-    public Builder setNumberValue(double value) {
-      kindCase_ = 2;
-      kind_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a double value.
-     * </pre>
-     *
-     * <code>double number_value = 2;</code>
-     */
-    public Builder clearNumberValue() {
-      if (kindCase_ == 2) {
-        kindCase_ = 0;
-        kind_ = null;
-        onChanged();
-      }
-      return this;
-    }
-
-    /**
-     * <pre>
-     * Represents a string value.
-     * </pre>
-     *
-     * <code>string string_value = 3;</code>
-     */
-    public java.lang.String getStringValue() {
-      java.lang.Object ref = "";
-      if (kindCase_ == 3) {
-        ref = kind_;
-      }
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (kindCase_ == 3) {
-          kind_ = s;
-        }
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * Represents a string value.
-     * </pre>
-     *
-     * <code>string string_value = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getStringValueBytes() {
-      java.lang.Object ref = "";
-      if (kindCase_ == 3) {
-        ref = kind_;
-      }
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        if (kindCase_ == 3) {
-          kind_ = b;
-        }
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * Represents a string value.
-     * </pre>
-     *
-     * <code>string string_value = 3;</code>
-     */
-    public Builder setStringValue(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  kindCase_ = 3;
-      kind_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a string value.
-     * </pre>
-     *
-     * <code>string string_value = 3;</code>
-     */
-    public Builder clearStringValue() {
-      if (kindCase_ == 3) {
-        kindCase_ = 0;
-        kind_ = null;
-        onChanged();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a string value.
-     * </pre>
-     *
-     * <code>string string_value = 3;</code>
-     */
-    public Builder setStringValueBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      kindCase_ = 3;
-      kind_ = value;
-      onChanged();
-      return this;
-    }
-
-    /**
-     * <pre>
-     * Represents a boolean value.
-     * </pre>
-     *
-     * <code>bool bool_value = 4;</code>
-     */
-    public boolean getBoolValue() {
-      if (kindCase_ == 4) {
-        return (java.lang.Boolean) kind_;
-      }
-      return false;
-    }
-    /**
-     * <pre>
-     * Represents a boolean value.
-     * </pre>
-     *
-     * <code>bool bool_value = 4;</code>
-     */
-    public Builder setBoolValue(boolean value) {
-      kindCase_ = 4;
-      kind_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a boolean value.
-     * </pre>
-     *
-     * <code>bool bool_value = 4;</code>
-     */
-    public Builder clearBoolValue() {
-      if (kindCase_ == 4) {
-        kindCase_ = 0;
-        kind_ = null;
-        onChanged();
-      }
-      return this;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct, org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.StructOrBuilder> structValueBuilder_;
-    /**
-     * <pre>
-     * Represents a structured value.
-     * </pre>
-     *
-     * <code>.google.protobuf.Struct struct_value = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct getStructValue() {
-      if (structValueBuilder_ == null) {
-        if (kindCase_ == 5) {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_;
-        }
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance();
-      } else {
-        if (kindCase_ == 5) {
-          return structValueBuilder_.getMessage();
-        }
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance();
-      }
-    }
-    /**
-     * <pre>
-     * Represents a structured value.
-     * </pre>
-     *
-     * <code>.google.protobuf.Struct struct_value = 5;</code>
-     */
-    public Builder setStructValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct value) {
-      if (structValueBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        kind_ = value;
-        onChanged();
-      } else {
-        structValueBuilder_.setMessage(value);
-      }
-      kindCase_ = 5;
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a structured value.
-     * </pre>
-     *
-     * <code>.google.protobuf.Struct struct_value = 5;</code>
-     */
-    public Builder setStructValue(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.Builder builderForValue) {
-      if (structValueBuilder_ == null) {
-        kind_ = builderForValue.build();
-        onChanged();
-      } else {
-        structValueBuilder_.setMessage(builderForValue.build());
-      }
-      kindCase_ = 5;
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a structured value.
-     * </pre>
-     *
-     * <code>.google.protobuf.Struct struct_value = 5;</code>
-     */
-    public Builder mergeStructValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct value) {
-      if (structValueBuilder_ == null) {
-        if (kindCase_ == 5 &&
-            kind_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance()) {
-          kind_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.newBuilder((org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_)
-              .mergeFrom(value).buildPartial();
-        } else {
-          kind_ = value;
-        }
-        onChanged();
-      } else {
-        if (kindCase_ == 5) {
-          structValueBuilder_.mergeFrom(value);
-        }
-        structValueBuilder_.setMessage(value);
-      }
-      kindCase_ = 5;
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a structured value.
-     * </pre>
-     *
-     * <code>.google.protobuf.Struct struct_value = 5;</code>
-     */
-    public Builder clearStructValue() {
-      if (structValueBuilder_ == null) {
-        if (kindCase_ == 5) {
-          kindCase_ = 0;
-          kind_ = null;
-          onChanged();
-        }
-      } else {
-        if (kindCase_ == 5) {
-          kindCase_ = 0;
-          kind_ = null;
-        }
-        structValueBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a structured value.
-     * </pre>
-     *
-     * <code>.google.protobuf.Struct struct_value = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.Builder getStructValueBuilder() {
-      return getStructValueFieldBuilder().getBuilder();
-    }
-    /**
-     * <pre>
-     * Represents a structured value.
-     * </pre>
-     *
-     * <code>.google.protobuf.Struct struct_value = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.StructOrBuilder getStructValueOrBuilder() {
-      if ((kindCase_ == 5) && (structValueBuilder_ != null)) {
-        return structValueBuilder_.getMessageOrBuilder();
-      } else {
-        if (kindCase_ == 5) {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_;
-        }
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance();
-      }
-    }
-    /**
-     * <pre>
-     * Represents a structured value.
-     * </pre>
-     *
-     * <code>.google.protobuf.Struct struct_value = 5;</code>
-     */
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct, org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.StructOrBuilder> 
-        getStructValueFieldBuilder() {
-      if (structValueBuilder_ == null) {
-        if (!(kindCase_ == 5)) {
-          kind_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance();
-        }
-        structValueBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct, org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.StructOrBuilder>(
-                (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) kind_,
-                getParentForChildren(),
-                isClean());
-        kind_ = null;
-      }
-      kindCase_ = 5;
-      onChanged();;
-      return structValueBuilder_;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue, org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValueOrBuilder> listValueBuilder_;
-    /**
-     * <pre>
-     * Represents a repeated `Value`.
-     * </pre>
-     *
-     * <code>.google.protobuf.ListValue list_value = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue getListValue() {
-      if (listValueBuilder_ == null) {
-        if (kindCase_ == 6) {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_;
-        }
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance();
-      } else {
-        if (kindCase_ == 6) {
-          return listValueBuilder_.getMessage();
-        }
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance();
-      }
-    }
-    /**
-     * <pre>
-     * Represents a repeated `Value`.
-     * </pre>
-     *
-     * <code>.google.protobuf.ListValue list_value = 6;</code>
-     */
-    public Builder setListValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue value) {
-      if (listValueBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        kind_ = value;
-        onChanged();
-      } else {
-        listValueBuilder_.setMessage(value);
-      }
-      kindCase_ = 6;
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a repeated `Value`.
-     * </pre>
-     *
-     * <code>.google.protobuf.ListValue list_value = 6;</code>
-     */
-    public Builder setListValue(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.Builder builderForValue) {
-      if (listValueBuilder_ == null) {
-        kind_ = builderForValue.build();
-        onChanged();
-      } else {
-        listValueBuilder_.setMessage(builderForValue.build());
-      }
-      kindCase_ = 6;
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a repeated `Value`.
-     * </pre>
-     *
-     * <code>.google.protobuf.ListValue list_value = 6;</code>
-     */
-    public Builder mergeListValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue value) {
-      if (listValueBuilder_ == null) {
-        if (kindCase_ == 6 &&
-            kind_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance()) {
-          kind_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.newBuilder((org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_)
-              .mergeFrom(value).buildPartial();
-        } else {
-          kind_ = value;
-        }
-        onChanged();
-      } else {
-        if (kindCase_ == 6) {
-          listValueBuilder_.mergeFrom(value);
-        }
-        listValueBuilder_.setMessage(value);
-      }
-      kindCase_ = 6;
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a repeated `Value`.
-     * </pre>
-     *
-     * <code>.google.protobuf.ListValue list_value = 6;</code>
-     */
-    public Builder clearListValue() {
-      if (listValueBuilder_ == null) {
-        if (kindCase_ == 6) {
-          kindCase_ = 0;
-          kind_ = null;
-          onChanged();
-        }
-      } else {
-        if (kindCase_ == 6) {
-          kindCase_ = 0;
-          kind_ = null;
-        }
-        listValueBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents a repeated `Value`.
-     * </pre>
-     *
-     * <code>.google.protobuf.ListValue list_value = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.Builder getListValueBuilder() {
-      return getListValueFieldBuilder().getBuilder();
-    }
-    /**
-     * <pre>
-     * Represents a repeated `Value`.
-     * </pre>
-     *
-     * <code>.google.protobuf.ListValue list_value = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValueOrBuilder getListValueOrBuilder() {
-      if ((kindCase_ == 6) && (listValueBuilder_ != null)) {
-        return listValueBuilder_.getMessageOrBuilder();
-      } else {
-        if (kindCase_ == 6) {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_;
-        }
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance();
-      }
-    }
-    /**
-     * <pre>
-     * Represents a repeated `Value`.
-     * </pre>
-     *
-     * <code>.google.protobuf.ListValue list_value = 6;</code>
-     */
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue, org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValueOrBuilder> 
-        getListValueFieldBuilder() {
-      if (listValueBuilder_ == null) {
-        if (!(kindCase_ == 6)) {
-          kind_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.getDefaultInstance();
-        }
-        listValueBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue, org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValueOrBuilder>(
-                (org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue) kind_,
-                getParentForChildren(),
-                isClean());
-        kind_ = null;
-      }
-      kindCase_ = 6;
-      onChanged();;
-      return listValueBuilder_;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Value)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Value)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Value DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Value();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Value>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Value>() {
-    public Value parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Value(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Value> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Value> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ValueOrBuilder.java
deleted file mode 100644
index 32d390f..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ValueOrBuilder.java
+++ /dev/null
@@ -1,98 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/struct.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface ValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Value)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * Represents a null value.
-   * </pre>
-   *
-   * <code>.google.protobuf.NullValue null_value = 1;</code>
-   */
-  int getNullValueValue();
-  /**
-   * <pre>
-   * Represents a null value.
-   * </pre>
-   *
-   * <code>.google.protobuf.NullValue null_value = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.NullValue getNullValue();
-
-  /**
-   * <pre>
-   * Represents a double value.
-   * </pre>
-   *
-   * <code>double number_value = 2;</code>
-   */
-  double getNumberValue();
-
-  /**
-   * <pre>
-   * Represents a string value.
-   * </pre>
-   *
-   * <code>string string_value = 3;</code>
-   */
-  java.lang.String getStringValue();
-  /**
-   * <pre>
-   * Represents a string value.
-   * </pre>
-   *
-   * <code>string string_value = 3;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getStringValueBytes();
-
-  /**
-   * <pre>
-   * Represents a boolean value.
-   * </pre>
-   *
-   * <code>bool bool_value = 4;</code>
-   */
-  boolean getBoolValue();
-
-  /**
-   * <pre>
-   * Represents a structured value.
-   * </pre>
-   *
-   * <code>.google.protobuf.Struct struct_value = 5;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct getStructValue();
-  /**
-   * <pre>
-   * Represents a structured value.
-   * </pre>
-   *
-   * <code>.google.protobuf.Struct struct_value = 5;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.StructOrBuilder getStructValueOrBuilder();
-
-  /**
-   * <pre>
-   * Represents a repeated `Value`.
-   * </pre>
-   *
-   * <code>.google.protobuf.ListValue list_value = 6;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValue getListValue();
-  /**
-   * <pre>
-   * Represents a repeated `Value`.
-   * </pre>
-   *
-   * <code>.google.protobuf.ListValue list_value = 6;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ListValueOrBuilder getListValueOrBuilder();
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.KindCase getKindCase();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/WireFormat.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/WireFormat.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/WireFormat.java
deleted file mode 100644
index 84d62f1..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/WireFormat.java
+++ /dev/null
@@ -1,260 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-
-/**
- * This class is used internally by the Protocol Buffer library and generated
- * message implementations.  It is public only because those generated messages
- * do not reside in the {@code protobuf} package.  Others should not use this
- * class directly.
- *
- * This class contains constants and helper functions useful for dealing with
- * the Protocol Buffer wire format.
- *
- * @author kenton@google.com Kenton Varda
- */
-public final class WireFormat {
-  // Do not allow instantiation.
-  private WireFormat() {}
-
-  static final int FIXED_32_SIZE = 4;
-  static final int FIXED_64_SIZE = 8;
-  static final int MAX_VARINT_SIZE = 10;
-
-  public static final int WIRETYPE_VARINT           = 0;
-  public static final int WIRETYPE_FIXED64          = 1;
-  public static final int WIRETYPE_LENGTH_DELIMITED = 2;
-  public static final int WIRETYPE_START_GROUP      = 3;
-  public static final int WIRETYPE_END_GROUP        = 4;
-  public static final int WIRETYPE_FIXED32          = 5;
-
-  static final int TAG_TYPE_BITS = 3;
-  static final int TAG_TYPE_MASK = (1 << TAG_TYPE_BITS) - 1;
-
-  /** Given a tag value, determines the wire type (the lower 3 bits). */
-  public static int getTagWireType(final int tag) {
-    return tag & TAG_TYPE_MASK;
-  }
-
-  /** Given a tag value, determines the field number (the upper 29 bits). */
-  public static int getTagFieldNumber(final int tag) {
-    return tag >>> TAG_TYPE_BITS;
-  }
-
-  /** Makes a tag value given a field number and wire type. */
-  static int makeTag(final int fieldNumber, final int wireType) {
-    return (fieldNumber << TAG_TYPE_BITS) | wireType;
-  }
-
-  /**
-   * Lite equivalent to {@link Descriptors.FieldDescriptor.JavaType}.  This is
-   * only here to support the lite runtime and should not be used by users.
-   */
-  public enum JavaType {
-    INT(0),
-    LONG(0L),
-    FLOAT(0F),
-    DOUBLE(0D),
-    BOOLEAN(false),
-    STRING(""),
-    BYTE_STRING(ByteString.EMPTY),
-    ENUM(null),
-    MESSAGE(null);
-
-    JavaType(final Object defaultDefault) {
-      this.defaultDefault = defaultDefault;
-    }
-
-    /**
-     * The default default value for fields of this type, if it's a primitive
-     * type.
-     */
-    Object getDefaultDefault() {
-      return defaultDefault;
-    }
-
-    private final Object defaultDefault;
-  }
-
-  /**
-   * Lite equivalent to {@link Descriptors.FieldDescriptor.Type}.  This is
-   * only here to support the lite runtime and should not be used by users.
-   */
-  public enum FieldType {
-    DOUBLE  (JavaType.DOUBLE     , WIRETYPE_FIXED64         ),
-    FLOAT   (JavaType.FLOAT      , WIRETYPE_FIXED32         ),
-    INT64   (JavaType.LONG       , WIRETYPE_VARINT          ),
-    UINT64  (JavaType.LONG       , WIRETYPE_VARINT          ),
-    INT32   (JavaType.INT        , WIRETYPE_VARINT          ),
-    FIXED64 (JavaType.LONG       , WIRETYPE_FIXED64         ),
-    FIXED32 (JavaType.INT        , WIRETYPE_FIXED32         ),
-    BOOL    (JavaType.BOOLEAN    , WIRETYPE_VARINT          ),
-    STRING  (JavaType.STRING     , WIRETYPE_LENGTH_DELIMITED) {
-      @Override
-      public boolean isPackable() {
-        return false; }
-    },
-    GROUP   (JavaType.MESSAGE    , WIRETYPE_START_GROUP     ) {
-      @Override
-      public boolean isPackable() {
-        return false; }
-    },
-    MESSAGE (JavaType.MESSAGE    , WIRETYPE_LENGTH_DELIMITED) {
-      @Override
-      public boolean isPackable() {
-        return false; }
-    },
-    BYTES   (JavaType.BYTE_STRING, WIRETYPE_LENGTH_DELIMITED) {
-      @Override
-      public boolean isPackable() {
-        return false; }
-    },
-    UINT32  (JavaType.INT        , WIRETYPE_VARINT          ),
-    ENUM    (JavaType.ENUM       , WIRETYPE_VARINT          ),
-    SFIXED32(JavaType.INT        , WIRETYPE_FIXED32         ),
-    SFIXED64(JavaType.LONG       , WIRETYPE_FIXED64         ),
-    SINT32  (JavaType.INT        , WIRETYPE_VARINT          ),
-    SINT64  (JavaType.LONG       , WIRETYPE_VARINT          );
-
-    FieldType(final JavaType javaType, final int wireType) {
-      this.javaType = javaType;
-      this.wireType = wireType;
-    }
-
-    private final JavaType javaType;
-    private final int wireType;
-
-    public JavaType getJavaType() { return javaType; }
-    public int getWireType() { return wireType; }
-
-    public boolean isPackable() { return true; }
-  }
-
-  // Field numbers for fields in MessageSet wire format.
-  static final int MESSAGE_SET_ITEM    = 1;
-  static final int MESSAGE_SET_TYPE_ID = 2;
-  static final int MESSAGE_SET_MESSAGE = 3;
-
-  // Tag numbers.
-  static final int MESSAGE_SET_ITEM_TAG =
-    makeTag(MESSAGE_SET_ITEM, WIRETYPE_START_GROUP);
-  static final int MESSAGE_SET_ITEM_END_TAG =
-    makeTag(MESSAGE_SET_ITEM, WIRETYPE_END_GROUP);
-  static final int MESSAGE_SET_TYPE_ID_TAG =
-    makeTag(MESSAGE_SET_TYPE_ID, WIRETYPE_VARINT);
-  static final int MESSAGE_SET_MESSAGE_TAG =
-    makeTag(MESSAGE_SET_MESSAGE, WIRETYPE_LENGTH_DELIMITED);
-
-  /**
-   * Validation level for handling incoming string field data which potentially
-   * contain non-UTF8 bytes.
-   */
-  enum Utf8Validation {
-    /** Eagerly parses to String; silently accepts invalid UTF8 bytes. */
-    LOOSE {
-      @Override
-      Object readString(CodedInputStream input) throws IOException {
-        return input.readString();
-      }
-    },
-    /** Eagerly parses to String; throws an IOException on invalid bytes. */
-    STRICT {
-      @Override
-      Object readString(CodedInputStream input) throws IOException {
-        return input.readStringRequireUtf8();
-      }
-    },
-    /** Keep data as ByteString; validation/conversion to String is lazy. */
-    LAZY {
-      @Override
-      Object readString(CodedInputStream input) throws IOException {
-        return input.readBytes();
-      }
-    };
-
-    /** Read a string field from the input with the proper UTF8 validation. */
-    abstract Object readString(CodedInputStream input) throws IOException;
-  }
-
-  /**
-   * Read a field of any primitive type for immutable messages from a
-   * CodedInputStream. Enums, groups, and embedded messages are not handled by
-   * this method.
-   *
-   * @param input The stream from which to read.
-   * @param type Declared type of the field.
-   * @param utf8Validation Different string UTF8 validation level for handling
-   *                       string fields.
-   * @return An object representing the field's value, of the exact
-   *         type which would be returned by
-   *         {@link Message#getField(Descriptors.FieldDescriptor)} for
-   *         this field.
-   */
-  static Object readPrimitiveField(
-      CodedInputStream input,
-      FieldType type,
-      Utf8Validation utf8Validation) throws IOException {
-    switch (type) {
-      case DOUBLE  : return input.readDouble  ();
-      case FLOAT   : return input.readFloat   ();
-      case INT64   : return input.readInt64   ();
-      case UINT64  : return input.readUInt64  ();
-      case INT32   : return input.readInt32   ();
-      case FIXED64 : return input.readFixed64 ();
-      case FIXED32 : return input.readFixed32 ();
-      case BOOL    : return input.readBool    ();
-      case BYTES   : return input.readBytes   ();
-      case UINT32  : return input.readUInt32  ();
-      case SFIXED32: return input.readSFixed32();
-      case SFIXED64: return input.readSFixed64();
-      case SINT32  : return input.readSInt32  ();
-      case SINT64  : return input.readSInt64  ();
-
-      case STRING  : return utf8Validation.readString(input);
-      case GROUP:
-        throw new IllegalArgumentException(
-          "readPrimitiveField() cannot handle nested groups.");
-      case MESSAGE:
-        throw new IllegalArgumentException(
-          "readPrimitiveField() cannot handle embedded messages.");
-      case ENUM:
-        // We don't handle enums because we don't know what to do if the
-        // value is not recognized.
-        throw new IllegalArgumentException(
-          "readPrimitiveField() cannot handle enums.");
-    }
-
-    throw new RuntimeException(
-      "There is no way to get here, but the compiler thinks otherwise.");
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/WrappersProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/WrappersProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/WrappersProto.java
deleted file mode 100644
index d3e9452..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/WrappersProto.java
+++ /dev/null
@@ -1,153 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class WrappersProto {
-  private WrappersProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_DoubleValue_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_DoubleValue_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_FloatValue_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_FloatValue_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Int64Value_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Int64Value_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_UInt64Value_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_UInt64Value_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Int32Value_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Int32Value_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_UInt32Value_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_UInt32Value_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_BoolValue_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_BoolValue_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_StringValue_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_StringValue_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_BytesValue_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_BytesValue_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\036google/protobuf/wrappers.proto\022\017google" +
-      ".protobuf\"\034\n\013DoubleValue\022\r\n\005value\030\001 \001(\001\"" +
-      "\033\n\nFloatValue\022\r\n\005value\030\001 \001(\002\"\033\n\nInt64Val" +
-      "ue\022\r\n\005value\030\001 \001(\003\"\034\n\013UInt64Value\022\r\n\005valu" +
-      "e\030\001 \001(\004\"\033\n\nInt32Value\022\r\n\005value\030\001 \001(\005\"\034\n\013" +
-      "UInt32Value\022\r\n\005value\030\001 \001(\r\"\032\n\tBoolValue\022" +
-      "\r\n\005value\030\001 \001(\010\"\034\n\013StringValue\022\r\n\005value\030\001" +
-      " \001(\t\"\033\n\nBytesValue\022\r\n\005value\030\001 \001(\014B|\n\023com" +
-      ".google.protobufB\rWrappersProtoP\001Z*githu" +
-      "b.com/golang/protobuf/ptypes/wrappers\370\001\001",
-      "\242\002\003GPB\252\002\036Google.Protobuf.WellKnownTypesb" +
-      "\006proto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_google_protobuf_DoubleValue_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_DoubleValue_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_DoubleValue_descriptor,
-        new java.lang.String[] { "Value", });
-    internal_static_google_protobuf_FloatValue_descriptor =
-      getDescriptor().getMessageTypes().get(1);
-    internal_static_google_protobuf_FloatValue_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_FloatValue_descriptor,
-        new java.lang.String[] { "Value", });
-    internal_static_google_protobuf_Int64Value_descriptor =
-      getDescriptor().getMessageTypes().get(2);
-    internal_static_google_protobuf_Int64Value_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Int64Value_descriptor,
-        new java.lang.String[] { "Value", });
-    internal_static_google_protobuf_UInt64Value_descriptor =
-      getDescriptor().getMessageTypes().get(3);
-    internal_static_google_protobuf_UInt64Value_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_UInt64Value_descriptor,
-        new java.lang.String[] { "Value", });
-    internal_static_google_protobuf_Int32Value_descriptor =
-      getDescriptor().getMessageTypes().get(4);
-    internal_static_google_protobuf_Int32Value_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Int32Value_descriptor,
-        new java.lang.String[] { "Value", });
-    internal_static_google_protobuf_UInt32Value_descriptor =
-      getDescriptor().getMessageTypes().get(5);
-    internal_static_google_protobuf_UInt32Value_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_UInt32Value_descriptor,
-        new java.lang.String[] { "Value", });
-    internal_static_google_protobuf_BoolValue_descriptor =
-      getDescriptor().getMessageTypes().get(6);
-    internal_static_google_protobuf_BoolValue_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_BoolValue_descriptor,
-        new java.lang.String[] { "Value", });
-    internal_static_google_protobuf_StringValue_descriptor =
-      getDescriptor().getMessageTypes().get(7);
-    internal_static_google_protobuf_StringValue_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_StringValue_descriptor,
-        new java.lang.String[] { "Value", });
-    internal_static_google_protobuf_BytesValue_descriptor =
-      getDescriptor().getMessageTypes().get(8);
-    internal_static_google_protobuf_BytesValue_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_BytesValue_descriptor,
-        new java.lang.String[] { "Value", });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[25/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessageLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessageLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessageLite.java
deleted file mode 100644
index 584eba1..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/GeneratedMessageLite.java
+++ /dev/null
@@ -1,2419 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.LimitedInputStream;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageLite.EqualsVisitor.NotEqualsException;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.BooleanList;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.DoubleList;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.FloatList;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.IntList;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.LongList;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.ProtobufList;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.FieldType;
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.ObjectStreamException;
-import java.io.Serializable;
-import java.lang.reflect.InvocationTargetException;
-import java.lang.reflect.Method;
-import java.util.ArrayList;
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-
-/**
- * Lite version of {@link GeneratedMessage}.
- *
- * @author kenton@google.com Kenton Varda
- */
-public abstract class GeneratedMessageLite<
-    MessageType extends GeneratedMessageLite<MessageType, BuilderType>,
-    BuilderType extends GeneratedMessageLite.Builder<MessageType, BuilderType>>
-        extends AbstractMessageLite<MessageType, BuilderType> {
-
-  /** For use by generated code only. Lazily initialized to reduce allocations. */
-  protected UnknownFieldSetLite unknownFields = UnknownFieldSetLite.getDefaultInstance();
-
-  /** For use by generated code only.  */
-  protected int memoizedSerializedSize = -1;
-
-  @Override
-  @SuppressWarnings("unchecked") // Guaranteed by runtime.
-  public final Parser<MessageType> getParserForType() {
-    return (Parser<MessageType>) dynamicMethod(MethodToInvoke.GET_PARSER);
-  }
-
-  @Override
-  @SuppressWarnings("unchecked") // Guaranteed by runtime.
-  public final MessageType getDefaultInstanceForType() {
-    return (MessageType) dynamicMethod(MethodToInvoke.GET_DEFAULT_INSTANCE);
-  }
-
-  @Override
-  @SuppressWarnings("unchecked") // Guaranteed by runtime.
-  public final BuilderType newBuilderForType() {
-    return (BuilderType) dynamicMethod(MethodToInvoke.NEW_BUILDER);
-  }
-
-  /**
-   * A reflective toString function. This is primarily intended as a developer aid, while keeping
-   * binary size down. The first line of the {@code toString()} representation includes a commented
-   * version of {@code super.toString()} to act as an indicator that this should not be relied on
-   * for comparisons.
-   * <p>
-   * NOTE: This method relies on the field getter methods not being stripped or renamed by proguard.
-   * If they are, the fields will not be included in the returned string representation.
-   * <p>
-   * NOTE: This implementation is liable to change in the future, and should not be relied on in
-   * code.
-   */
-  @Override
-  public String toString() {
-    return MessageLiteToString.toString(this, super.toString());
-  }
-
-  @SuppressWarnings("unchecked") // Guaranteed by runtime
-  @Override
-  public int hashCode() {
-    if (memoizedHashCode == 0) {
-      HashCodeVisitor visitor = new HashCodeVisitor();
-      visit(visitor, (MessageType) this);
-      memoizedHashCode = visitor.hashCode;
-    }
-    return memoizedHashCode;
-  }
-
-  @SuppressWarnings("unchecked") // Guaranteed by runtime
-  int hashCode(HashCodeVisitor visitor) {
-    if (memoizedHashCode == 0) {
-      int inProgressHashCode = visitor.hashCode;
-      visitor.hashCode = 0;
-      visit(visitor, (MessageType) this);
-      memoizedHashCode = visitor.hashCode;
-      visitor.hashCode = inProgressHashCode;
-    }
-    return memoizedHashCode;
-  }
-
-  @SuppressWarnings("unchecked") // Guaranteed by isInstance + runtime
-  @Override
-  public boolean equals(Object other) {
-    if (this == other) {
-      return true;
-    }
-
-    if (!getDefaultInstanceForType().getClass().isInstance(other)) {
-      return false;
-    }
-
-    try {
-      visit(EqualsVisitor.INSTANCE, (MessageType) other);
-    } catch (NotEqualsException e) {
-      return false;
-    }
-    return true;
-  }
-
-  /**
-   * Same as {@link #equals(Object)} but throws {@code NotEqualsException}.
-   */
-  @SuppressWarnings("unchecked") // Guaranteed by isInstance + runtime
-  boolean equals(EqualsVisitor visitor, MessageLite other) {
-    if (this == other) {
-      return true;
-    }
-
-    if (!getDefaultInstanceForType().getClass().isInstance(other)) {
-      return false;
-    }
-
-    visit(visitor, (MessageType) other);
-    return true;
-  }
-
-  // The general strategy for unknown fields is to use an UnknownFieldSetLite that is treated as
-  // mutable during the parsing constructor and immutable after. This allows us to avoid
-  // any unnecessary intermediary allocations while reducing the generated code size.
-
-  /**
-   * Lazily initializes unknown fields.
-   */
-  private final void ensureUnknownFieldsInitialized() {
-    if (unknownFields == UnknownFieldSetLite.getDefaultInstance()) {
-      unknownFields = UnknownFieldSetLite.newInstance();
-    }
-  }
-
-  /**
-   * Called by subclasses to parse an unknown field. For use by generated code only.
-   *
-   * @return {@code true} unless the tag is an end-group tag.
-   */
-  protected boolean parseUnknownField(int tag, CodedInputStream input) throws IOException {
-    // This will avoid the allocation of unknown fields when a group tag is encountered.
-    if (WireFormat.getTagWireType(tag) == WireFormat.WIRETYPE_END_GROUP) {
-      return false;
-    }
-
-    ensureUnknownFieldsInitialized();
-    return unknownFields.mergeFieldFrom(tag, input);
-  }
-
-  /**
-   * Called by subclasses to parse an unknown field. For use by generated code only.
-   */
-  protected void mergeVarintField(int tag, int value) {
-    ensureUnknownFieldsInitialized();
-    unknownFields.mergeVarintField(tag, value);
-  }
-
-  /**
-   * Called by subclasses to parse an unknown field. For use by generated code only.
-   */
-  protected void mergeLengthDelimitedField(int fieldNumber, ByteString value) {
-    ensureUnknownFieldsInitialized();
-    unknownFields.mergeLengthDelimitedField(fieldNumber, value);
-  }
-
-  /**
-   * Called by subclasses to complete parsing. For use by generated code only.
-   */
-  protected void makeImmutable() {
-    dynamicMethod(MethodToInvoke.MAKE_IMMUTABLE);
-
-    unknownFields.makeImmutable();
-  }
-
-  @Override
-  public final boolean isInitialized() {
-    return dynamicMethod(MethodToInvoke.IS_INITIALIZED, Boolean.TRUE) != null;
-  }
-
-  @Override
-  public final BuilderType toBuilder() {
-    BuilderType builder = (BuilderType) dynamicMethod(MethodToInvoke.NEW_BUILDER);
-    builder.mergeFrom((MessageType) this);
-    return builder;
-  }
-
-  /**
-   * Defines which method path to invoke in {@link GeneratedMessageLite
-   * #dynamicMethod(MethodToInvoke, Object...)}.
-   * <p>
-   * For use by generated code only.
-   */
-  public static enum MethodToInvoke {
-    // Rely on/modify instance state
-    IS_INITIALIZED,
-    VISIT,
-    MERGE_FROM_STREAM,
-    MAKE_IMMUTABLE,
-
-    // Rely on static state
-    NEW_MUTABLE_INSTANCE,
-    NEW_BUILDER,
-    GET_DEFAULT_INSTANCE,
-    GET_PARSER;
-  }
-
-  /**
-   * A method that implements different types of operations described in {@link MethodToInvoke}.
-   * Theses different kinds of operations are required to implement message-level operations for
-   * builders in the runtime. This method bundles those operations to reduce the generated methods
-   * count.
-   * <ul>
-   * <li>{@code MERGE_FROM_STREAM} is parameterized with an {@link CodedInputStream} and
-   * {@link ExtensionRegistryLite}. It consumes the input stream, parsing the contents into the
-   * returned protocol buffer. If parsing throws an {@link InvalidProtocolBufferException}, the
-   * implementation wraps it in a RuntimeException.
-   * <li>{@code NEW_INSTANCE} returns a new instance of the protocol buffer that has not yet been
-   * made immutable. See {@code MAKE_IMMUTABLE}.
-   * <li>{@code IS_INITIALIZED} is parameterized with a {@code Boolean} detailing whether to
-   * memoize. It returns {@code null} for false and the default instance for true. We optionally
-   * memoize to support the Builder case, where memoization is not desired.
-   * <li>{@code NEW_BUILDER} returns a {@code BuilderType} instance.
-   * <li>{@code VISIT} is parameterized with a {@code Visitor} and a {@code MessageType} and
-   * recursively iterates through the fields side by side between this and the instance.
-   * <li>{@code MAKE_IMMUTABLE} sets all internal fields to an immutable state.
-   * </ul>
-   * This method, plus the implementation of the Builder, enables the Builder class to be proguarded
-   * away entirely on Android.
-   * <p>
-   * For use by generated code only.
-   */
-  protected abstract Object dynamicMethod(MethodToInvoke method, Object arg0, Object arg1);
-
-  /**
-   * Same as {@link #dynamicMethod(MethodToInvoke, Object, Object)} with {@code null} padding.
-   */
-  protected Object dynamicMethod(MethodToInvoke method, Object arg0) {
-    return dynamicMethod(method, arg0, null);
-  }
-
-  /**
-   * Same as {@link #dynamicMethod(MethodToInvoke, Object, Object)} with {@code null} padding.
-   */
-  protected Object dynamicMethod(MethodToInvoke method) {
-    return dynamicMethod(method, null, null);
-  }
-
-  void visit(Visitor visitor, MessageType other) {
-    dynamicMethod(MethodToInvoke.VISIT, visitor, other);
-    unknownFields = visitor.visitUnknownFields(unknownFields, other.unknownFields);
-  }
-
-  /**
-   * Merge some unknown fields into the {@link UnknownFieldSetLite} for this
-   * message.
-   *
-   * <p>For use by generated code only.
-   */
-  protected final void mergeUnknownFields(UnknownFieldSetLite unknownFields) {
-    this.unknownFields = UnknownFieldSetLite.mutableCopyOf(this.unknownFields, unknownFields);
-  }
-
-  @SuppressWarnings("unchecked")
-  public abstract static class Builder<
-      MessageType extends GeneratedMessageLite<MessageType, BuilderType>,
-      BuilderType extends Builder<MessageType, BuilderType>>
-          extends AbstractMessageLite.Builder<MessageType, BuilderType> {
-
-    private final MessageType defaultInstance;
-    protected MessageType instance;
-    protected boolean isBuilt;
-
-    protected Builder(MessageType defaultInstance) {
-      this.defaultInstance = defaultInstance;
-      this.instance =
-          (MessageType) defaultInstance.dynamicMethod(MethodToInvoke.NEW_MUTABLE_INSTANCE);
-      isBuilt = false;
-    }
-
-    /**
-     * Called before any method that would mutate the builder to ensure that it correctly copies
-     * any state before the write happens to preserve immutability guarantees.
-     */
-    protected void copyOnWrite() {
-      if (isBuilt) {
-        MessageType newInstance =
-            (MessageType) instance.dynamicMethod(MethodToInvoke.NEW_MUTABLE_INSTANCE);
-        newInstance.visit(MergeFromVisitor.INSTANCE, instance);
-        instance = newInstance;
-        isBuilt = false;
-      }
-    }
-
-    @Override
-    public final boolean isInitialized() {
-      return GeneratedMessageLite.isInitialized(instance, false /* shouldMemoize */);
-    }
-
-    @Override
-    public final BuilderType clear() {
-      // No need to copy on write since we're dropping the instance anyways.
-      instance = (MessageType) instance.dynamicMethod(MethodToInvoke.NEW_MUTABLE_INSTANCE);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public BuilderType clone() {
-      BuilderType builder =
-          (BuilderType) getDefaultInstanceForType().newBuilderForType();
-      builder.mergeFrom(buildPartial());
-      return builder;
-    }
-
-    @Override
-    public MessageType buildPartial() {
-      if (isBuilt) {
-        return instance;
-      }
-
-      instance.makeImmutable();
-
-      isBuilt = true;
-      return instance;
-    }
-
-    @Override
-    public final MessageType build() {
-      MessageType result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    @Override
-    protected BuilderType internalMergeFrom(MessageType message) {
-      return mergeFrom(message);
-    }
-
-    /** All subclasses implement this. */
-    public BuilderType mergeFrom(MessageType message) {
-      copyOnWrite();
-      instance.visit(MergeFromVisitor.INSTANCE, message);
-      return (BuilderType) this;
-    }
-
-    @Override
-    public MessageType getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    @Override
-    public BuilderType mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      copyOnWrite();
-      try {
-        instance.dynamicMethod(MethodToInvoke.MERGE_FROM_STREAM, input, extensionRegistry);
-      } catch (RuntimeException e) {
-        if (e.getCause() instanceof IOException) {
-          throw (IOException) e.getCause();
-        }
-        throw e;
-      }
-      return (BuilderType) this;
-    }
-  }
-
-
-  // =================================================================
-  // Extensions-related stuff
-
-  /**
-   * Lite equivalent of {@link org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessage.ExtendableMessageOrBuilder}.
-   */
-  public interface ExtendableMessageOrBuilder<
-      MessageType extends ExtendableMessage<MessageType, BuilderType>,
-      BuilderType extends ExtendableBuilder<MessageType, BuilderType>>
-          extends MessageLiteOrBuilder {
-
-    /** Check if a singular extension is present. */
-    <Type> boolean hasExtension(
-        ExtensionLite<MessageType, Type> extension);
-
-    /** Get the number of elements in a repeated extension. */
-    <Type> int getExtensionCount(
-        ExtensionLite<MessageType, List<Type>> extension);
-
-    /** Get the value of an extension. */
-    <Type> Type getExtension(ExtensionLite<MessageType, Type> extension);
-
-    /** Get one element of a repeated extension. */
-    <Type> Type getExtension(
-        ExtensionLite<MessageType, List<Type>> extension,
-        int index);
-  }
-
-  /**
-   * Lite equivalent of {@link GeneratedMessage.ExtendableMessage}.
-   */
-  public abstract static class ExtendableMessage<
-        MessageType extends ExtendableMessage<MessageType, BuilderType>,
-        BuilderType extends ExtendableBuilder<MessageType, BuilderType>>
-            extends GeneratedMessageLite<MessageType, BuilderType>
-            implements ExtendableMessageOrBuilder<MessageType, BuilderType> {
-
-    /**
-     * Represents the set of extensions on this message. For use by generated
-     * code only.
-     */
-    protected FieldSet<ExtensionDescriptor> extensions = FieldSet.newFieldSet();
-
-    protected final void mergeExtensionFields(final MessageType other) {
-      if (extensions.isImmutable()) {
-        extensions = extensions.clone();
-      }
-      extensions.mergeFrom(((ExtendableMessage) other).extensions);
-    }
-
-    @Override
-    final void visit(Visitor visitor, MessageType other) {
-      super.visit(visitor, other);
-      extensions = visitor.visitExtensions(extensions, other.extensions);
-    }
-
-    /**
-     * Parse an unknown field or an extension. For use by generated code only.
-     *
-     * <p>For use by generated code only.
-     *
-     * @return {@code true} unless the tag is an end-group tag.
-     */
-    protected <MessageType extends MessageLite> boolean parseUnknownField(
-        MessageType defaultInstance,
-        CodedInputStream input,
-        ExtensionRegistryLite extensionRegistry,
-        int tag) throws IOException {
-      int fieldNumber = WireFormat.getTagFieldNumber(tag);
-
-      // TODO(dweis): How much bytecode would be saved by not requiring the generated code to
-      //     provide the default instance?
-      GeneratedExtension<MessageType, ?> extension = extensionRegistry.findLiteExtensionByNumber(
-          defaultInstance, fieldNumber);
-
-      return parseExtension(input, extensionRegistry, extension, tag, fieldNumber);
-    }
-
-    private boolean parseExtension(
-        CodedInputStream input,
-        ExtensionRegistryLite extensionRegistry,
-        GeneratedExtension<?, ?> extension,
-        int tag,
-        int fieldNumber)
-        throws IOException {
-      int wireType = WireFormat.getTagWireType(tag);
-      boolean unknown = false;
-      boolean packed = false;
-      if (extension == null) {
-        unknown = true;  // Unknown field.
-      } else if (wireType == FieldSet.getWireFormatForFieldType(
-                   extension.descriptor.getLiteType(),
-                   false  /* isPacked */)) {
-        packed = false;  // Normal, unpacked value.
-      } else if (extension.descriptor.isRepeated &&
-                 extension.descriptor.type.isPackable() &&
-                 wireType == FieldSet.getWireFormatForFieldType(
-                   extension.descriptor.getLiteType(),
-                   true  /* isPacked */)) {
-        packed = true;  // Packed value.
-      } else {
-        unknown = true;  // Wrong wire type.
-      }
-
-      if (unknown) {  // Unknown field or wrong wire type.  Skip.
-        return parseUnknownField(tag, input);
-      }
-      
-      if (packed) {
-        int length = input.readRawVarint32();
-        int limit = input.pushLimit(length);
-        if (extension.descriptor.getLiteType() == WireFormat.FieldType.ENUM) {
-          while (input.getBytesUntilLimit() > 0) {
-            int rawValue = input.readEnum();
-            Object value =
-                extension.descriptor.getEnumType().findValueByNumber(rawValue);
-            if (value == null) {
-              // If the number isn't recognized as a valid value for this
-              // enum, drop it (don't even add it to unknownFields).
-              return true;
-            }
-            extensions.addRepeatedField(extension.descriptor,
-                                        extension.singularToFieldSetType(value));
-          }
-        } else {
-          while (input.getBytesUntilLimit() > 0) {
-            Object value =
-                FieldSet.readPrimitiveField(input,
-                                            extension.descriptor.getLiteType(),
-                                            /*checkUtf8=*/ false);
-            extensions.addRepeatedField(extension.descriptor, value);
-          }
-        }
-        input.popLimit(limit);
-      } else {
-        Object value;
-        switch (extension.descriptor.getLiteJavaType()) {
-          case MESSAGE: {
-            MessageLite.Builder subBuilder = null;
-            if (!extension.descriptor.isRepeated()) {
-              MessageLite existingValue =
-                  (MessageLite) extensions.getField(extension.descriptor);
-              if (existingValue != null) {
-                subBuilder = existingValue.toBuilder();
-              }
-            }
-            if (subBuilder == null) {
-              subBuilder = extension.getMessageDefaultInstance()
-                  .newBuilderForType();
-            }
-            if (extension.descriptor.getLiteType() ==
-                WireFormat.FieldType.GROUP) {
-              input.readGroup(extension.getNumber(),
-                              subBuilder, extensionRegistry);
-            } else {
-              input.readMessage(subBuilder, extensionRegistry);
-            }
-            value = subBuilder.build();
-            break;
-          }
-          case ENUM:
-            int rawValue = input.readEnum();
-            value = extension.descriptor.getEnumType()
-                             .findValueByNumber(rawValue);
-            // If the number isn't recognized as a valid value for this enum,
-            // write it to unknown fields object.
-            if (value == null) {
-              mergeVarintField(fieldNumber, rawValue);
-              return true;
-            }
-            break;
-          default:
-            value = FieldSet.readPrimitiveField(input,
-                extension.descriptor.getLiteType(),
-                /*checkUtf8=*/ false);
-            break;
-        }
-
-        if (extension.descriptor.isRepeated()) {
-          extensions.addRepeatedField(extension.descriptor,
-                                      extension.singularToFieldSetType(value));
-        } else {
-          extensions.setField(extension.descriptor,
-                              extension.singularToFieldSetType(value));
-        }
-      }
-      return true;
-    }
-    
-    /**
-     * Parse an unknown field or an extension. For use by generated code only.
-     *
-     * <p>For use by generated code only.
-     *
-     * @return {@code true} unless the tag is an end-group tag.
-     */
-    protected <MessageType extends MessageLite> boolean parseUnknownFieldAsMessageSet(
-        MessageType defaultInstance,
-        CodedInputStream input,
-        ExtensionRegistryLite extensionRegistry,
-        int tag)
-        throws IOException {
-
-      if (tag == WireFormat.MESSAGE_SET_ITEM_TAG) {
-        mergeMessageSetExtensionFromCodedStream(defaultInstance, input, extensionRegistry);
-        return true;
-      }
-
-      // TODO(dweis): Do we really want to support non message set wire format in message sets?
-      // Full runtime does... So we do for now.
-      int wireType = WireFormat.getTagWireType(tag);
-      if (wireType == WireFormat.WIRETYPE_LENGTH_DELIMITED) {
-        return parseUnknownField(defaultInstance, input, extensionRegistry, tag);
-      } else {
-        // TODO(dweis): Should we throw on invalid input? Full runtime does not...
-        return input.skipField(tag);
-      }
-    }
-
-    /**
-     * Merges the message set from the input stream; requires message set wire format.
-     * 
-     * @param defaultInstance the default instance of the containing message we are parsing in
-     * @param input the stream to parse from
-     * @param extensionRegistry the registry to use when parsing
-     */
-    private <MessageType extends MessageLite> void mergeMessageSetExtensionFromCodedStream(
-        MessageType defaultInstance,
-        CodedInputStream input,
-        ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      // The wire format for MessageSet is:
-      //   message MessageSet {
-      //     repeated group Item = 1 {
-      //       required int32 typeId = 2;
-      //       required bytes message = 3;
-      //     }
-      //   }
-      // "typeId" is the extension's field number.  The extension can only be
-      // a message type, where "message" contains the encoded bytes of that
-      // message.
-      //
-      // In practice, we will probably never see a MessageSet item in which
-      // the message appears before the type ID, or where either field does not
-      // appear exactly once.  However, in theory such cases are valid, so we
-      // should be prepared to accept them.
-
-      int typeId = 0;
-      ByteString rawBytes = null; // If we encounter "message" before "typeId"
-      GeneratedExtension<?, ?> extension = null;
-
-      // Read bytes from input, if we get it's type first then parse it eagerly,
-      // otherwise we store the raw bytes in a local variable.
-      while (true) {
-        final int tag = input.readTag();
-        if (tag == 0) {
-          break;
-        }
-
-        if (tag == WireFormat.MESSAGE_SET_TYPE_ID_TAG) {
-          typeId = input.readUInt32();
-          if (typeId != 0) {
-            extension = extensionRegistry.findLiteExtensionByNumber(defaultInstance, typeId);
-          }
-
-        } else if (tag == WireFormat.MESSAGE_SET_MESSAGE_TAG) {
-          if (typeId != 0) {
-            if (extension != null) {
-              // We already know the type, so we can parse directly from the
-              // input with no copying.  Hooray!
-              eagerlyMergeMessageSetExtension(input, extension, extensionRegistry, typeId);
-              rawBytes = null;
-              continue;
-            }
-          }
-          // We haven't seen a type ID yet or we want parse message lazily.
-          rawBytes = input.readBytes();
-
-        } else { // Unknown tag. Skip it.
-          if (!input.skipField(tag)) {
-            break; // End of group
-          }
-        }
-      }
-      input.checkLastTagWas(WireFormat.MESSAGE_SET_ITEM_END_TAG);
-
-      // Process the raw bytes.
-      if (rawBytes != null && typeId != 0) { // Zero is not a valid type ID.
-        if (extension != null) { // We known the type
-          mergeMessageSetExtensionFromBytes(rawBytes, extensionRegistry, extension);
-        } else { // We don't know how to parse this. Ignore it.
-          if (rawBytes != null) {
-            mergeLengthDelimitedField(typeId, rawBytes);
-          }
-        }
-      }
-    }
-
-    private void eagerlyMergeMessageSetExtension(
-        CodedInputStream input,
-        GeneratedExtension<?, ?> extension,
-        ExtensionRegistryLite extensionRegistry,
-        int typeId)
-        throws IOException {
-      int fieldNumber = typeId;
-      int tag = WireFormat.makeTag(typeId, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-      parseExtension(input, extensionRegistry, extension, tag, fieldNumber);
-    }
-
-    private void mergeMessageSetExtensionFromBytes(
-        ByteString rawBytes,
-        ExtensionRegistryLite extensionRegistry,
-        GeneratedExtension<?, ?> extension)
-        throws IOException {
-      MessageLite.Builder subBuilder = null;
-      MessageLite existingValue = (MessageLite) extensions.getField(extension.descriptor);
-      if (existingValue != null) {
-        subBuilder = existingValue.toBuilder();
-      }
-      if (subBuilder == null) {
-        subBuilder = extension.getMessageDefaultInstance().newBuilderForType();
-      }
-      rawBytes.newCodedInput().readMessage(subBuilder, extensionRegistry);
-      MessageLite value = subBuilder.build();
-
-      extensions.setField(extension.descriptor, extension.singularToFieldSetType(value));
-    }
-
-    private void verifyExtensionContainingType(
-        final GeneratedExtension<MessageType, ?> extension) {
-      if (extension.getContainingTypeDefaultInstance() !=
-          getDefaultInstanceForType()) {
-        // This can only happen if someone uses unchecked operations.
-        throw new IllegalArgumentException(
-          "This extension is for a different message type.  Please make " +
-          "sure that you are not suppressing any generics type warnings.");
-      }
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final ExtensionLite<MessageType, Type> extension) {
-      GeneratedExtension<MessageType, Type> extensionLite =
-          checkIsLite(extension);
-
-      verifyExtensionContainingType(extensionLite);
-      return extensions.hasField(extensionLite.descriptor);
-    }
-
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final ExtensionLite<MessageType, List<Type>> extension) {
-      GeneratedExtension<MessageType, List<Type>> extensionLite =
-          checkIsLite(extension);
-
-      verifyExtensionContainingType(extensionLite);
-      return extensions.getRepeatedFieldCount(extensionLite.descriptor);
-    }
-
-    /** Get the value of an extension. */
-    @Override
-    @SuppressWarnings("unchecked")
-    public final <Type> Type getExtension(final ExtensionLite<MessageType, Type> extension) {
-      GeneratedExtension<MessageType, Type> extensionLite =
-          checkIsLite(extension);
-
-      verifyExtensionContainingType(extensionLite);
-      final Object value = extensions.getField(extensionLite.descriptor);
-      if (value == null) {
-        return extensionLite.defaultValue;
-      } else {
-        return (Type) extensionLite.fromFieldSetType(value);
-      }
-    }
-
-    /** Get one element of a repeated extension. */
-    @Override
-    @SuppressWarnings("unchecked")
-    public final <Type> Type getExtension(
-        final ExtensionLite<MessageType, List<Type>> extension, final int index) {
-      GeneratedExtension<MessageType, List<Type>> extensionLite =
-          checkIsLite(extension);
-
-      verifyExtensionContainingType(extensionLite);
-      return (Type) extensionLite.singularFromFieldSetType(
-          extensions.getRepeatedField(extensionLite.descriptor, index));
-    }
-
-    /** Called by subclasses to check if all extensions are initialized. */
-    protected boolean extensionsAreInitialized() {
-      return extensions.isInitialized();
-    }
-
-    @Override
-    protected final void makeImmutable() {
-      super.makeImmutable();
-
-      extensions.makeImmutable();
-    }
-
-    /**
-     * Used by subclasses to serialize extensions.  Extension ranges may be
-     * interleaved with field numbers, but we must write them in canonical
-     * (sorted by field number) order.  ExtensionWriter helps us write
-     * individual ranges of extensions at once.
-     */
-    protected class ExtensionWriter {
-      // Imagine how much simpler this code would be if Java iterators had
-      // a way to get the next element without advancing the iterator.
-
-      private final Iterator<Map.Entry<ExtensionDescriptor, Object>> iter =
-            extensions.iterator();
-      private Map.Entry<ExtensionDescriptor, Object> next;
-      private final boolean messageSetWireFormat;
-
-      private ExtensionWriter(boolean messageSetWireFormat) {
-        if (iter.hasNext()) {
-          next = iter.next();
-        }
-        this.messageSetWireFormat = messageSetWireFormat;
-      }
-
-      public void writeUntil(final int end, final CodedOutputStream output)
-                             throws IOException {
-        while (next != null && next.getKey().getNumber() < end) {
-          ExtensionDescriptor extension = next.getKey();
-          if (messageSetWireFormat && extension.getLiteJavaType() ==
-                  WireFormat.JavaType.MESSAGE &&
-              !extension.isRepeated()) {
-            output.writeMessageSetExtension(extension.getNumber(),
-                                            (MessageLite) next.getValue());
-          } else {
-            FieldSet.writeField(extension, next.getValue(), output);
-          }
-          if (iter.hasNext()) {
-            next = iter.next();
-          } else {
-            next = null;
-          }
-        }
-      }
-    }
-
-    protected ExtensionWriter newExtensionWriter() {
-      return new ExtensionWriter(false);
-    }
-    protected ExtensionWriter newMessageSetExtensionWriter() {
-      return new ExtensionWriter(true);
-    }
-
-    /** Called by subclasses to compute the size of extensions. */
-    protected int extensionsSerializedSize() {
-      return extensions.getSerializedSize();
-    }
-    protected int extensionsSerializedSizeAsMessageSet() {
-      return extensions.getMessageSetSerializedSize();
-    }
-  }
-
-  /**
-   * Lite equivalent of {@link GeneratedMessage.ExtendableBuilder}.
-   */
-  @SuppressWarnings("unchecked")
-  public abstract static class ExtendableBuilder<
-        MessageType extends ExtendableMessage<MessageType, BuilderType>,
-        BuilderType extends ExtendableBuilder<MessageType, BuilderType>>
-      extends Builder<MessageType, BuilderType>
-      implements ExtendableMessageOrBuilder<MessageType, BuilderType> {
-    protected ExtendableBuilder(MessageType defaultInstance) {
-      super(defaultInstance);
-
-      // TODO(dweis): This is kind of an unnecessary clone since we construct a
-      //     new instance in the parent constructor which makes the extensions
-      //     immutable. This extra allocation shouldn't matter in practice
-      //     though.
-      instance.extensions = instance.extensions.clone();
-    }
-
-    // For immutable message conversion.
-    void internalSetExtensionSet(FieldSet<ExtensionDescriptor> extensions) {
-      copyOnWrite();
-      instance.extensions = extensions;
-    }
-
-    @Override
-    protected void copyOnWrite() {
-      if (!isBuilt) {
-        return;
-      }
-
-      super.copyOnWrite();
-      instance.extensions = instance.extensions.clone();
-    }
-
-    @Override
-    public final MessageType buildPartial() {
-      if (isBuilt) {
-        return instance;
-      }
-
-      instance.extensions.makeImmutable();
-      return super.buildPartial();
-    }
-
-    private void verifyExtensionContainingType(
-        final GeneratedExtension<MessageType, ?> extension) {
-      if (extension.getContainingTypeDefaultInstance() !=
-          getDefaultInstanceForType()) {
-        // This can only happen if someone uses unchecked operations.
-        throw new IllegalArgumentException(
-          "This extension is for a different message type.  Please make " +
-          "sure that you are not suppressing any generics type warnings.");
-      }
-    }
-
-    /** Check if a singular extension is present. */
-    @Override
-    public final <Type> boolean hasExtension(final ExtensionLite<MessageType, Type> extension) {
-      return instance.hasExtension(extension);
-    }
-
-    /** Get the number of elements in a repeated extension. */
-    @Override
-    public final <Type> int getExtensionCount(
-        final ExtensionLite<MessageType, List<Type>> extension) {
-      return instance.getExtensionCount(extension);
-    }
-
-    /** Get the value of an extension. */
-    @Override
-    @SuppressWarnings("unchecked")
-    public final <Type> Type getExtension(final ExtensionLite<MessageType, Type> extension) {
-      return instance.getExtension(extension);
-    }
-
-    /** Get one element of a repeated extension. */
-    @Override
-    @SuppressWarnings("unchecked")
-    public final <Type> Type getExtension(
-        final ExtensionLite<MessageType, List<Type>> extension, final int index) {
-      return instance.getExtension(extension, index);
-    }
-
-    /** Set the value of an extension. */
-    public final <Type> BuilderType setExtension(
-        final ExtensionLite<MessageType, Type> extension,
-        final Type value) {
-      GeneratedExtension<MessageType, Type> extensionLite =
-          checkIsLite(extension);
-
-      verifyExtensionContainingType(extensionLite);
-      copyOnWrite();
-      instance.extensions.setField(extensionLite.descriptor, extensionLite.toFieldSetType(value));
-      return (BuilderType) this;
-    }
-
-    /** Set the value of one element of a repeated extension. */
-    public final <Type> BuilderType setExtension(
-        final ExtensionLite<MessageType, List<Type>> extension,
-        final int index, final Type value) {
-      GeneratedExtension<MessageType, List<Type>> extensionLite =
-          checkIsLite(extension);
-
-      verifyExtensionContainingType(extensionLite);
-      copyOnWrite();
-      instance.extensions.setRepeatedField(
-          extensionLite.descriptor, index, extensionLite.singularToFieldSetType(value));
-      return (BuilderType) this;
-    }
-
-    /** Append a value to a repeated extension. */
-    public final <Type> BuilderType addExtension(
-        final ExtensionLite<MessageType, List<Type>> extension,
-        final Type value) {
-      GeneratedExtension<MessageType, List<Type>> extensionLite =
-          checkIsLite(extension);
-
-      verifyExtensionContainingType(extensionLite);
-      copyOnWrite();
-      instance.extensions.addRepeatedField(
-          extensionLite.descriptor, extensionLite.singularToFieldSetType(value));
-      return (BuilderType) this;
-    }
-
-    /** Clear an extension. */
-    public final <Type> BuilderType clearExtension(
-        final ExtensionLite<MessageType, ?> extension) {
-      GeneratedExtension<MessageType, ?> extensionLite = checkIsLite(extension);
-
-      verifyExtensionContainingType(extensionLite);
-      copyOnWrite();
-      instance.extensions.clearField(extensionLite.descriptor);
-      return (BuilderType) this;
-    }
-  }
-
-  // -----------------------------------------------------------------
-
-  /** For use by generated code only. */
-  public static <ContainingType extends MessageLite, Type>
-      GeneratedExtension<ContainingType, Type>
-          newSingularGeneratedExtension(
-              final ContainingType containingTypeDefaultInstance,
-              final Type defaultValue,
-              final MessageLite messageDefaultInstance,
-              final Internal.EnumLiteMap<?> enumTypeMap,
-              final int number,
-              final WireFormat.FieldType type,
-              final Class singularType) {
-    return new GeneratedExtension<ContainingType, Type>(
-        containingTypeDefaultInstance,
-        defaultValue,
-        messageDefaultInstance,
-        new ExtensionDescriptor(enumTypeMap, number, type,
-                                false /* isRepeated */,
-                                false /* isPacked */),
-        singularType);
-  }
-
-  /** For use by generated code only. */
-  public static <ContainingType extends MessageLite, Type>
-      GeneratedExtension<ContainingType, Type>
-          newRepeatedGeneratedExtension(
-              final ContainingType containingTypeDefaultInstance,
-              final MessageLite messageDefaultInstance,
-              final Internal.EnumLiteMap<?> enumTypeMap,
-              final int number,
-              final WireFormat.FieldType type,
-              final boolean isPacked,
-              final Class singularType) {
-    @SuppressWarnings("unchecked")  // Subclasses ensure Type is a List
-    Type emptyList = (Type) Collections.emptyList();
-    return new GeneratedExtension<ContainingType, Type>(
-        containingTypeDefaultInstance,
-        emptyList,
-        messageDefaultInstance,
-        new ExtensionDescriptor(
-            enumTypeMap, number, type, true /* isRepeated */, isPacked),
-        singularType);
-  }
-
-  static final class ExtensionDescriptor
-      implements FieldSet.FieldDescriptorLite<
-        ExtensionDescriptor> {
-    ExtensionDescriptor(
-        final Internal.EnumLiteMap<?> enumTypeMap,
-        final int number,
-        final WireFormat.FieldType type,
-        final boolean isRepeated,
-        final boolean isPacked) {
-      this.enumTypeMap = enumTypeMap;
-      this.number = number;
-      this.type = type;
-      this.isRepeated = isRepeated;
-      this.isPacked = isPacked;
-    }
-
-    final Internal.EnumLiteMap<?> enumTypeMap;
-    final int number;
-    final WireFormat.FieldType type;
-    final boolean isRepeated;
-    final boolean isPacked;
-
-    @Override
-    public int getNumber() {
-      return number;
-    }
-
-    @Override
-    public WireFormat.FieldType getLiteType() {
-      return type;
-    }
-
-    @Override
-    public WireFormat.JavaType getLiteJavaType() {
-      return type.getJavaType();
-    }
-
-    @Override
-    public boolean isRepeated() {
-      return isRepeated;
-    }
-
-    @Override
-    public boolean isPacked() {
-      return isPacked;
-    }
-
-    @Override
-    public Internal.EnumLiteMap<?> getEnumType() {
-      return enumTypeMap;
-    }
-
-    @Override
-    @SuppressWarnings("unchecked")
-    public MessageLite.Builder internalMergeFrom(MessageLite.Builder to, MessageLite from) {
-      return ((Builder) to).mergeFrom((GeneratedMessageLite) from);
-    }
-
-
-    @Override
-    public int compareTo(ExtensionDescriptor other) {
-      return number - other.number;
-    }
-  }
-
-  // =================================================================
-
-  /** Calls Class.getMethod and throws a RuntimeException if it fails. */
-  @SuppressWarnings("unchecked")
-  static Method getMethodOrDie(Class clazz, String name, Class... params) {
-    try {
-      return clazz.getMethod(name, params);
-    } catch (NoSuchMethodException e) {
-      throw new RuntimeException(
-        "Generated message class \"" + clazz.getName() +
-        "\" missing method \"" + name + "\".", e);
-    }
-  }
-
-  /** Calls invoke and throws a RuntimeException if it fails. */
-  static Object invokeOrDie(Method method, Object object, Object... params) {
-    try {
-      return method.invoke(object, params);
-    } catch (IllegalAccessException e) {
-      throw new RuntimeException(
-        "Couldn't use Java reflection to implement protocol message " +
-        "reflection.", e);
-    } catch (InvocationTargetException e) {
-      final Throwable cause = e.getCause();
-      if (cause instanceof RuntimeException) {
-        throw (RuntimeException) cause;
-      } else if (cause instanceof Error) {
-        throw (Error) cause;
-      } else {
-        throw new RuntimeException(
-          "Unexpected exception thrown by generated accessor method.", cause);
-      }
-    }
-  }
-
-  /**
-   * Lite equivalent to {@link GeneratedMessage.GeneratedExtension}.
-   *
-   * Users should ignore the contents of this class and only use objects of
-   * this type as parameters to extension accessors and ExtensionRegistry.add().
-   */
-  public static class GeneratedExtension<
-      ContainingType extends MessageLite, Type>
-          extends ExtensionLite<ContainingType, Type> {
-
-    /**
-     * Create a new instance with the given parameters.
-     *
-     * The last parameter {@code singularType} is only needed for enum types.
-     * We store integer values for enum types in a {@link ExtendableMessage}
-     * and use Java reflection to convert an integer value back into a concrete
-     * enum object.
-     */
-    GeneratedExtension(
-        final ContainingType containingTypeDefaultInstance,
-        final Type defaultValue,
-        final MessageLite messageDefaultInstance,
-        final ExtensionDescriptor descriptor,
-        final Class singularType) {
-      // Defensive checks to verify the correct initialization order of
-      // GeneratedExtensions and their related GeneratedMessages.
-      if (containingTypeDefaultInstance == null) {
-        throw new IllegalArgumentException(
-            "Null containingTypeDefaultInstance");
-      }
-      if (descriptor.getLiteType() == WireFormat.FieldType.MESSAGE &&
-          messageDefaultInstance == null) {
-        throw new IllegalArgumentException(
-            "Null messageDefaultInstance");
-      }
-      this.containingTypeDefaultInstance = containingTypeDefaultInstance;
-      this.defaultValue = defaultValue;
-      this.messageDefaultInstance = messageDefaultInstance;
-      this.descriptor = descriptor;
-    }
-
-    final ContainingType containingTypeDefaultInstance;
-    final Type defaultValue;
-    final MessageLite messageDefaultInstance;
-    final ExtensionDescriptor descriptor;
-
-    /**
-     * Default instance of the type being extended, used to identify that type.
-     */
-    public ContainingType getContainingTypeDefaultInstance() {
-      return containingTypeDefaultInstance;
-    }
-
-    /** Get the field number. */
-    @Override
-    public int getNumber() {
-      return descriptor.getNumber();
-    }
-
-
-    /**
-     * If the extension is an embedded message or group, returns the default
-     * instance of the message.
-     */
-    @Override
-    public MessageLite getMessageDefaultInstance() {
-      return messageDefaultInstance;
-    }
-
-    @SuppressWarnings("unchecked")
-    Object fromFieldSetType(final Object value) {
-      if (descriptor.isRepeated()) {
-        if (descriptor.getLiteJavaType() == WireFormat.JavaType.ENUM) {
-          final List result = new ArrayList();
-          for (final Object element : (List) value) {
-            result.add(singularFromFieldSetType(element));
-          }
-          return result;
-        } else {
-          return value;
-        }
-      } else {
-        return singularFromFieldSetType(value);
-      }
-    }
-
-    Object singularFromFieldSetType(final Object value) {
-      if (descriptor.getLiteJavaType() == WireFormat.JavaType.ENUM) {
-        return descriptor.enumTypeMap.findValueByNumber((Integer) value);
-      } else {
-        return value;
-      }
-    }
-
-    @SuppressWarnings("unchecked")
-    Object toFieldSetType(final Object value) {
-      if (descriptor.isRepeated()) {
-        if (descriptor.getLiteJavaType() == WireFormat.JavaType.ENUM) {
-          final List result = new ArrayList();
-          for (final Object element : (List) value) {
-            result.add(singularToFieldSetType(element));
-          }
-          return result;
-        } else {
-          return value;
-        }
-      } else {
-        return singularToFieldSetType(value);
-      }
-    }
-
-    Object singularToFieldSetType(final Object value) {
-      if (descriptor.getLiteJavaType() == WireFormat.JavaType.ENUM) {
-        return ((Internal.EnumLite) value).getNumber();
-      } else {
-        return value;
-      }
-    }
-
-    @Override
-    public FieldType getLiteType() {
-      return descriptor.getLiteType();
-    }
-
-    @Override
-    public boolean isRepeated() {
-      return descriptor.isRepeated;
-    }
-
-    @Override
-    public Type getDefaultValue() {
-      return defaultValue;
-    }
-  }
-
-  /**
-   * A serialized (serializable) form of the generated message.  Stores the
-   * message as a class name and a byte array.
-   */
-  protected static final class SerializedForm implements Serializable {
-
-    public static SerializedForm of(MessageLite message) {
-      return new SerializedForm(message);
-    }
-
-    private static final long serialVersionUID = 0L;
-
-    private final String messageClassName;
-    private final byte[] asBytes;
-
-    /**
-     * Creates the serialized form by calling {@link org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageLite#toByteArray}.
-     * @param regularForm the message to serialize
-     */
-    SerializedForm(MessageLite regularForm) {
-      messageClassName = regularForm.getClass().getName();
-      asBytes = regularForm.toByteArray();
-    }
-
-    /**
-     * When read from an ObjectInputStream, this method converts this object
-     * back to the regular form.  Part of Java's serialization magic.
-     * @return a GeneratedMessage of the type that was serialized
-     */
-    @SuppressWarnings("unchecked")
-    protected Object readResolve() throws ObjectStreamException {
-      try {
-        Class<?> messageClass = Class.forName(messageClassName);
-        java.lang.reflect.Field defaultInstanceField =
-            messageClass.getDeclaredField("DEFAULT_INSTANCE");
-        defaultInstanceField.setAccessible(true);
-        MessageLite defaultInstance = (MessageLite) defaultInstanceField.get(null);
-        return defaultInstance.newBuilderForType()
-            .mergeFrom(asBytes)
-            .buildPartial();
-      } catch (ClassNotFoundException e) {
-        throw new RuntimeException("Unable to find proto buffer class: " + messageClassName, e);
-      } catch (NoSuchFieldException e) {
-        return readResolveFallback();
-      } catch (SecurityException e) {
-        throw new RuntimeException("Unable to call DEFAULT_INSTANCE in " + messageClassName, e);
-      } catch (IllegalAccessException e) {
-        throw new RuntimeException("Unable to call parsePartialFrom", e);
-      } catch (InvalidProtocolBufferException e) {
-        throw new RuntimeException("Unable to understand proto buffer", e);
-      }
-    }
-
-    /**
-     * @deprecated from v3.0.0-beta-3+, for compatibility with v2.5.0 and v2.6.1 generated code.
-     */
-    @Deprecated
-    private Object readResolveFallback() throws ObjectStreamException {
-      try {
-        Class<?> messageClass = Class.forName(messageClassName);
-        java.lang.reflect.Field defaultInstanceField =
-            messageClass.getDeclaredField("defaultInstance");
-        defaultInstanceField.setAccessible(true);
-        MessageLite defaultInstance = (MessageLite) defaultInstanceField.get(null);
-        return defaultInstance.newBuilderForType()
-            .mergeFrom(asBytes)
-            .buildPartial();
-      } catch (ClassNotFoundException e) {
-        throw new RuntimeException("Unable to find proto buffer class: " + messageClassName, e);
-      } catch (NoSuchFieldException e) {
-        throw new RuntimeException("Unable to find defaultInstance in " + messageClassName, e);
-      } catch (SecurityException e) {
-        throw new RuntimeException("Unable to call defaultInstance in " + messageClassName, e);
-      } catch (IllegalAccessException e) {
-        throw new RuntimeException("Unable to call parsePartialFrom", e);
-      } catch (InvalidProtocolBufferException e) {
-        throw new RuntimeException("Unable to understand proto buffer", e);
-      }
-    }
-  }
-
-  /**
-   * Checks that the {@link Extension} is Lite and returns it as a
-   * {@link GeneratedExtension}.
-   */
-  private static <
-      MessageType extends ExtendableMessage<MessageType, BuilderType>,
-      BuilderType extends ExtendableBuilder<MessageType, BuilderType>,
-      T>
-    GeneratedExtension<MessageType, T> checkIsLite(
-        ExtensionLite<MessageType, T> extension) {
-    if (!extension.isLite()) {
-      throw new IllegalArgumentException("Expected a lite extension.");
-    }
-
-    return (GeneratedExtension<MessageType, T>) extension;
-  }
-
-  /**
-   * A static helper method for checking if a message is initialized, optionally memoizing.
-   * <p>
-   * For use by generated code only.
-   */
-  protected static final <T extends GeneratedMessageLite<T, ?>> boolean isInitialized(
-      T message, boolean shouldMemoize) {
-    return message.dynamicMethod(MethodToInvoke.IS_INITIALIZED, shouldMemoize) != null;
-  }
-
-  protected static final <T extends GeneratedMessageLite<T, ?>> void makeImmutable(T message) {
-    message.dynamicMethod(MethodToInvoke.MAKE_IMMUTABLE);
-  }
-
-  protected static IntList emptyIntList() {
-    return IntArrayList.emptyList();
-  }
-
-  protected static IntList mutableCopy(IntList list) {
-    int size = list.size();
-    return list.mutableCopyWithCapacity(
-        size == 0 ? AbstractProtobufList.DEFAULT_CAPACITY : size * 2);
-  }
-
-  protected static LongList emptyLongList() {
-    return LongArrayList.emptyList();
-  }
-
-  protected static LongList mutableCopy(LongList list) {
-    int size = list.size();
-    return list.mutableCopyWithCapacity(
-        size == 0 ? AbstractProtobufList.DEFAULT_CAPACITY : size * 2);
-  }
-
-  protected static FloatList emptyFloatList() {
-    return FloatArrayList.emptyList();
-  }
-
-  protected static FloatList mutableCopy(FloatList list) {
-    int size = list.size();
-    return list.mutableCopyWithCapacity(
-        size == 0 ? AbstractProtobufList.DEFAULT_CAPACITY : size * 2);
-  }
-
-  protected static DoubleList emptyDoubleList() {
-    return DoubleArrayList.emptyList();
-  }
-
-  protected static DoubleList mutableCopy(DoubleList list) {
-    int size = list.size();
-    return list.mutableCopyWithCapacity(
-        size == 0 ? AbstractProtobufList.DEFAULT_CAPACITY : size * 2);
-  }
-
-  protected static BooleanList emptyBooleanList() {
-    return BooleanArrayList.emptyList();
-  }
-
-  protected static BooleanList mutableCopy(BooleanList list) {
-    int size = list.size();
-    return list.mutableCopyWithCapacity(
-        size == 0 ? AbstractProtobufList.DEFAULT_CAPACITY : size * 2);
-  }
-
-  protected static <E> ProtobufList<E> emptyProtobufList() {
-    return ProtobufArrayList.emptyList();
-  }
-
-  protected static <E> ProtobufList<E> mutableCopy(ProtobufList<E> list) {
-    int size = list.size();
-    return list.mutableCopyWithCapacity(
-        size == 0 ? AbstractProtobufList.DEFAULT_CAPACITY : size * 2);
-  }
-
-  /**
-   * A {@link Parser} implementation that delegates to the default instance.
-   * <p>
-   * For use by generated code only.
-   */
-  protected static class DefaultInstanceBasedParser<T extends GeneratedMessageLite<T, ?>>
-      extends AbstractParser<T> {
-
-    private T defaultInstance;
-
-    public DefaultInstanceBasedParser(T defaultInstance) {
-      this.defaultInstance = defaultInstance;
-    }
-
-    @Override
-    public T parsePartialFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      return GeneratedMessageLite.parsePartialFrom(defaultInstance, input, extensionRegistry);
-    }
-  }
-
-  /**
-   * A static helper method for parsing a partial from input using the extension registry and the
-   * instance.
-   */
-  // TODO(dweis): Should this verify that the last tag was 0?
-  static <T extends GeneratedMessageLite<T, ?>> T parsePartialFrom(
-      T instance, CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-          throws InvalidProtocolBufferException {
-    @SuppressWarnings("unchecked") // Guaranteed by protoc
-    T result = (T) instance.dynamicMethod(MethodToInvoke.NEW_MUTABLE_INSTANCE);
-    try {
-      result.dynamicMethod(MethodToInvoke.MERGE_FROM_STREAM, input, extensionRegistry);
-      result.makeImmutable();
-    } catch (RuntimeException e) {
-      if (e.getCause() instanceof InvalidProtocolBufferException) {
-        throw (InvalidProtocolBufferException) e.getCause();
-      }
-      throw e;
-    }
-    return result;
-  }
-
-  protected static <T extends GeneratedMessageLite<T, ?>> T parsePartialFrom(
-      T defaultInstance,
-      CodedInputStream input)
-      throws InvalidProtocolBufferException {
-    return parsePartialFrom(defaultInstance, input, ExtensionRegistryLite.getEmptyRegistry());
-  }
-
-  /**
-   * Helper method to check if message is initialized.
-   *
-   * @throws InvalidProtocolBufferException if it is not initialized.
-   * @return The message to check.
-   */
-  private static <T extends GeneratedMessageLite<T, ?>> T checkMessageInitialized(T message)
-      throws InvalidProtocolBufferException {
-    if (message != null && !message.isInitialized()) {
-      throw message.newUninitializedMessageException()
-          .asInvalidProtocolBufferException()
-          .setUnfinishedMessage(message);
-    }
-    return message;
-  }
-
-  // Validates last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseFrom(
-      T defaultInstance, ByteString data)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parseFrom(defaultInstance, data, ExtensionRegistryLite.getEmptyRegistry()));
-  }
-
-  // Validates last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseFrom(
-      T defaultInstance, ByteString data, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(parsePartialFrom(defaultInstance, data, extensionRegistry));
-  }
-
-  // This is a special case since we want to verify that the last tag is 0. We assume we exhaust the
-  // ByteString.
-  private static <T extends GeneratedMessageLite<T, ?>> T parsePartialFrom(
-      T defaultInstance, ByteString data, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    T message;
-    try {
-      CodedInputStream input = data.newCodedInput();
-      message = parsePartialFrom(defaultInstance, input, extensionRegistry);
-      try {
-        input.checkLastTagWas(0);
-      } catch (InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(message);
-      }
-      return message;
-    } catch (InvalidProtocolBufferException e) {
-      throw e;
-    }
-  }
-
-  // This is a special case since we want to verify that the last tag is 0. We assume we exhaust the
-  // ByteString.
-  private static <T extends GeneratedMessageLite<T, ?>> T parsePartialFrom(
-      T defaultInstance, byte[] data, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    T message;
-    try {
-      CodedInputStream input = CodedInputStream.newInstance(data);
-      message = parsePartialFrom(defaultInstance, input, extensionRegistry);
-      try {
-        input.checkLastTagWas(0);
-      } catch (InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(message);
-      }
-      return message;
-    } catch (InvalidProtocolBufferException e) {
-      throw e;
-    }
-  }
-
-  // Validates last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseFrom(
-      T defaultInstance, byte[] data)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialFrom(defaultInstance, data, ExtensionRegistryLite.getEmptyRegistry()));
-  }
-
-  // Validates last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseFrom(
-      T defaultInstance, byte[] data, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(parsePartialFrom(defaultInstance, data, extensionRegistry));
-  }
-
-  // Does not validate last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseFrom(
-      T defaultInstance, InputStream input)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialFrom(defaultInstance, CodedInputStream.newInstance(input),
-            ExtensionRegistryLite.getEmptyRegistry()));
-  }
-
-  // Does not validate last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseFrom(
-      T defaultInstance, InputStream input, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialFrom(defaultInstance, CodedInputStream.newInstance(input), extensionRegistry));
-  }
-
-  // Does not validate last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseFrom(
-      T defaultInstance, CodedInputStream input)
-      throws InvalidProtocolBufferException {
-    return parseFrom(defaultInstance, input, ExtensionRegistryLite.getEmptyRegistry());
-  }
-
-  // Does not validate last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseFrom(
-      T defaultInstance, CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialFrom(defaultInstance, input, extensionRegistry));
-  }
-
-  // Validates last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseDelimitedFrom(
-      T defaultInstance, InputStream input)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialDelimitedFrom(defaultInstance, input,
-            ExtensionRegistryLite.getEmptyRegistry()));
-  }
-
-  // Validates last tag.
-  protected static <T extends GeneratedMessageLite<T, ?>> T parseDelimitedFrom(
-      T defaultInstance, InputStream input, ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    return checkMessageInitialized(
-        parsePartialDelimitedFrom(defaultInstance, input, extensionRegistry));
-  }
-
-  private static <T extends GeneratedMessageLite<T, ?>> T parsePartialDelimitedFrom(
-      T defaultInstance,
-      InputStream input,
-      ExtensionRegistryLite extensionRegistry)
-      throws InvalidProtocolBufferException {
-    int size;
-    try {
-      int firstByte = input.read();
-      if (firstByte == -1) {
-        return null;
-      }
-      size = CodedInputStream.readRawVarint32(firstByte, input);
-    } catch (IOException e) {
-      throw new InvalidProtocolBufferException(e.getMessage());
-    }
-    InputStream limitedInput = new LimitedInputStream(input, size);
-    CodedInputStream codedInput = CodedInputStream.newInstance(limitedInput);
-    T message = parsePartialFrom(defaultInstance, codedInput, extensionRegistry);
-    try {
-      codedInput.checkLastTagWas(0);
-    } catch (InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(message);
-    }
-    return message;
-  }
-
-  /**
-   * An abstract visitor that the generated code calls into that we use to implement various
-   * features. Fields that are not members of oneofs are always visited. Members of a oneof are only
-   * visited when they are the set oneof case value on the "other" proto. The visitOneofNotSet
-   * method is invoked if other's oneof case is not set.
-   */
-  protected interface Visitor {
-    boolean visitBoolean(boolean minePresent, boolean mine, boolean otherPresent, boolean other);
-    int visitInt(boolean minePresent, int mine, boolean otherPresent, int other);
-    double visitDouble(boolean minePresent, double mine, boolean otherPresent, double other);
-    float visitFloat(boolean minePresent, float mine, boolean otherPresent, float other);
-    long visitLong(boolean minePresent, long mine, boolean otherPresent, long other);
-    String visitString(boolean minePresent, String mine, boolean otherPresent, String other);
-    ByteString visitByteString(
-        boolean minePresent, ByteString mine, boolean otherPresent, ByteString other);
-
-    Object visitOneofBoolean(boolean minePresent, Object mine, Object other);
-    Object visitOneofInt(boolean minePresent, Object mine, Object other);
-    Object visitOneofDouble(boolean minePresent, Object mine, Object other);
-    Object visitOneofFloat(boolean minePresent, Object mine, Object other);
-    Object visitOneofLong(boolean minePresent, Object mine, Object other);
-    Object visitOneofString(boolean minePresent, Object mine, Object other);
-    Object visitOneofByteString(boolean minePresent, Object mine, Object other);
-    Object visitOneofLazyMessage(boolean minePresent, Object mine, Object other);
-    Object visitOneofMessage(boolean minePresent, Object mine, Object other);
-    void visitOneofNotSet(boolean minePresent);
-
-    /**
-     * Message fields use null sentinals.
-     */
-    <T extends MessageLite> T visitMessage(T mine, T other);
-    LazyFieldLite visitLazyMessage(LazyFieldLite mine, LazyFieldLite other);
-
-    <T> ProtobufList<T> visitList(ProtobufList<T> mine, ProtobufList<T> other);
-    BooleanList visitBooleanList(BooleanList mine, BooleanList other);
-    IntList visitIntList(IntList mine, IntList other);
-    DoubleList visitDoubleList(DoubleList mine, DoubleList other);
-    FloatList visitFloatList(FloatList mine, FloatList other);
-    LongList visitLongList(LongList mine, LongList other);
-    FieldSet<ExtensionDescriptor> visitExtensions(
-        FieldSet<ExtensionDescriptor> mine, FieldSet<ExtensionDescriptor> other);
-    UnknownFieldSetLite visitUnknownFields(UnknownFieldSetLite mine, UnknownFieldSetLite other);
-    <K, V> MapFieldLite<K, V> visitMap(MapFieldLite<K, V> mine, MapFieldLite<K, V> other);
-  }
-
-  /**
-   * Implements equals. Throws a {@link NotEqualsException} when not equal.
-   */
-  static class EqualsVisitor implements Visitor {
-
-    static final class NotEqualsException extends RuntimeException {}
-
-    static final EqualsVisitor INSTANCE = new EqualsVisitor();
-
-    static final NotEqualsException NOT_EQUALS = new NotEqualsException();
-
-    private EqualsVisitor() {}
-
-    @Override
-    public boolean visitBoolean(
-        boolean minePresent, boolean mine, boolean otherPresent, boolean other) {
-      if (minePresent != otherPresent || mine != other) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public int visitInt(boolean minePresent, int mine, boolean otherPresent, int other) {
-      if (minePresent != otherPresent || mine != other) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public double visitDouble(
-        boolean minePresent, double mine, boolean otherPresent, double other) {
-      if (minePresent != otherPresent || mine != other) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public float visitFloat(boolean minePresent, float mine, boolean otherPresent, float other) {
-      if (minePresent != otherPresent || mine != other) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public long visitLong(boolean minePresent, long mine, boolean otherPresent, long other) {
-      if (minePresent != otherPresent || mine != other) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public String visitString(
-        boolean minePresent, String mine, boolean otherPresent, String other) {
-      if (minePresent != otherPresent || !mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public ByteString visitByteString(
-        boolean minePresent, ByteString mine, boolean otherPresent, ByteString other) {
-      if (minePresent != otherPresent || !mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofBoolean(boolean minePresent, Object mine, Object other) {
-      if (minePresent && mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public Object visitOneofInt(boolean minePresent, Object mine, Object other) {
-      if (minePresent && mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public Object visitOneofDouble(boolean minePresent, Object mine, Object other) {
-      if (minePresent && mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public Object visitOneofFloat(boolean minePresent, Object mine, Object other) {
-      if (minePresent && mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public Object visitOneofLong(boolean minePresent, Object mine, Object other) {
-      if (minePresent && mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public Object visitOneofString(boolean minePresent, Object mine, Object other) {
-      if (minePresent && mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public Object visitOneofByteString(boolean minePresent, Object mine, Object other) {
-      if (minePresent && mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public Object visitOneofLazyMessage(boolean minePresent, Object mine, Object other) {
-      if (minePresent && mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public Object visitOneofMessage(boolean minePresent, Object mine, Object other) {
-      if (minePresent && ((GeneratedMessageLite<?, ?>) mine).equals(this, (MessageLite) other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public void visitOneofNotSet(boolean minePresent) {
-      if (minePresent) {
-        throw NOT_EQUALS;
-      }
-    }
-
-    @Override
-    public <T extends MessageLite> T visitMessage(T mine, T other) {
-      if (mine == null && other == null) {
-        return null;
-      }
-
-      if (mine == null || other == null) {
-        throw NOT_EQUALS;
-      }
-
-      ((GeneratedMessageLite<?, ?>) mine).equals(this, other);
-
-      return mine;
-    }
-
-    @Override
-    public LazyFieldLite visitLazyMessage(
-        LazyFieldLite mine, LazyFieldLite other) {
-      if (mine == null && other == null) {
-        return null;
-      }
-      if (mine == null || other == null) {
-        throw NOT_EQUALS;
-      }
-      if (mine.equals(other)) {
-        return mine;
-      }
-      throw NOT_EQUALS;
-    }
-
-    @Override
-    public <T> ProtobufList<T> visitList(ProtobufList<T> mine, ProtobufList<T> other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public BooleanList visitBooleanList(BooleanList mine, BooleanList other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public IntList visitIntList(IntList mine, IntList other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public DoubleList visitDoubleList(DoubleList mine, DoubleList other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public FloatList visitFloatList(FloatList mine, FloatList other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public LongList visitLongList(LongList mine, LongList other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public FieldSet<ExtensionDescriptor> visitExtensions(
-        FieldSet<ExtensionDescriptor> mine,
-        FieldSet<ExtensionDescriptor> other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public UnknownFieldSetLite visitUnknownFields(
-        UnknownFieldSetLite mine,
-        UnknownFieldSetLite other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-
-    @Override
-    public <K, V> MapFieldLite<K, V> visitMap(MapFieldLite<K, V> mine, MapFieldLite<K, V> other) {
-      if (!mine.equals(other)) {
-        throw NOT_EQUALS;
-      }
-      return mine;
-    }
-  }
-
-  /**
-   * Implements hashCode by accumulating state.
-   */
-  private static class HashCodeVisitor implements Visitor {
-
-    // The caller must ensure that the visitor is invoked parameterized with this and this such that
-    // other is this. This is required due to how oneof cases are handled. See the class comment
-    // on Visitor for more information.
-
-    private int hashCode = 0;
-
-    @Override
-    public boolean visitBoolean(
-        boolean minePresent, boolean mine, boolean otherPresent, boolean other) {
-      hashCode = (53 * hashCode) + Internal.hashBoolean(mine);
-      return mine;
-    }
-
-    @Override
-    public int visitInt(boolean minePresent, int mine, boolean otherPresent, int other) {
-      hashCode = (53 * hashCode) + mine;
-      return mine;
-    }
-
-    @Override
-    public double visitDouble(
-        boolean minePresent, double mine, boolean otherPresent, double other) {
-      hashCode = (53 * hashCode) + Internal.hashLong(Double.doubleToLongBits(mine));
-      return mine;
-    }
-
-    @Override
-    public float visitFloat(boolean minePresent, float mine, boolean otherPresent, float other) {
-      hashCode = (53 * hashCode) + Float.floatToIntBits(mine);
-      return mine;
-    }
-
-    @Override
-    public long visitLong(boolean minePresent, long mine, boolean otherPresent, long other) {
-      hashCode = (53 * hashCode) + Internal.hashLong(mine);
-      return mine;
-    }
-
-    @Override
-    public String visitString(
-        boolean minePresent, String mine, boolean otherPresent, String other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public ByteString visitByteString(
-        boolean minePresent, ByteString mine, boolean otherPresent, ByteString other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofBoolean(boolean minePresent, Object mine, Object other) {
-      hashCode = (53 * hashCode) + Internal.hashBoolean(((Boolean) mine));
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofInt(boolean minePresent, Object mine, Object other) {
-      hashCode = (53 * hashCode) + (Integer) mine;
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofDouble(boolean minePresent, Object mine, Object other) {
-      hashCode = (53 * hashCode) + Internal.hashLong(Double.doubleToLongBits((Double) mine));
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofFloat(boolean minePresent, Object mine, Object other) {
-      hashCode = (53 * hashCode) + Float.floatToIntBits((Float) mine);
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofLong(boolean minePresent, Object mine, Object other) {
-      hashCode = (53 * hashCode) + Internal.hashLong((Long) mine);
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofString(boolean minePresent, Object mine, Object other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofByteString(boolean minePresent, Object mine, Object other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofLazyMessage(boolean minePresent, Object mine, Object other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public Object visitOneofMessage(boolean minePresent, Object mine, Object other) {
-      return visitMessage((MessageLite) mine, (MessageLite) other);
-    }
-
-    @Override
-    public void visitOneofNotSet(boolean minePresent) {
-      if (minePresent) {
-        throw new IllegalStateException(); // Can't happen if other == this.
-      }
-    }
-
-    @Override
-    public <T extends MessageLite> T visitMessage(T mine, T other) {
-      final int protoHash;
-      if (mine != null) {
-        if (mine instanceof GeneratedMessageLite) {
-          protoHash = ((GeneratedMessageLite) mine).hashCode(this);
-        } else {
-          protoHash = mine.hashCode();
-        }
-      } else {
-        protoHash = 37;
-      }
-      hashCode = (53 * hashCode) + protoHash;
-      return mine;
-    }
-
-    @Override
-    public LazyFieldLite visitLazyMessage(LazyFieldLite mine, LazyFieldLite other) {
-      final int protoHash;
-      if (mine != null) {
-        protoHash = mine.hashCode();
-      } else {
-        protoHash = 37;
-      }
-      hashCode = (53 * hashCode) + protoHash;
-      return mine;
-    }
-
-    @Override
-    public <T> ProtobufList<T> visitList(ProtobufList<T> mine, ProtobufList<T> other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public BooleanList visitBooleanList(BooleanList mine, BooleanList other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public IntList visitIntList(IntList mine, IntList other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public DoubleList visitDoubleList(DoubleList mine, DoubleList other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public FloatList visitFloatList(FloatList mine, FloatList other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public LongList visitLongList(LongList mine, LongList other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public FieldSet<ExtensionDescriptor> visitExtensions(
-        FieldSet<ExtensionDescriptor> mine,
-        FieldSet<ExtensionDescriptor> other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public UnknownFieldSetLite visitUnknownFields(
-        UnknownFieldSetLite mine,
-        UnknownFieldSetLite other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-
-    @Override
-    public <K, V> MapFieldLite<K, V> visitMap(MapFieldLite<K, V> mine, MapFieldLite<K, V> other) {
-      hashCode = (53 * hashCode) + mine.hashCode();
-      return mine;
-    }
-  }
-
-  /**
-   * Implements field merging semantics over the visitor interface.
-   */
-  protected static class MergeFromVisitor implements Visitor {
-
-    public static final MergeFromVisitor INSTANCE = new MergeFromVisitor();
-
-    private MergeFromVisitor() {}
-
-    @Override
-    public boolean visitBoolean(
-        boolean minePresent, boolean mine, boolean otherPresent, boolean other) {
-      return otherPresent ? other : mine;
-    }
-
-    @Override
-    public int visitInt(boolean minePresent, int mine, boolean otherPresent, int other) {
-      return otherPresent ? other : mine;
-    }
-
-    @Override
-    public double visitDouble(
-        boolean minePresent, double mine, boolean otherPresent, double other) {
-      return otherPresent ? other : mine;
-    }
-
-    @Override
-    public float visitFloat(boolean minePresent, float mine, boolean otherPresent, float other) {
-      return otherPresent ? other : mine;
-    }
-
-    @Override
-    public long visitLong(boolean minePresent, long mine, boolean otherPresent, long other) {
-      return otherPresent ? other : mine;
-    }
-
-    @Override
-    public String visitString(
-        boolean minePresent, String mine, boolean otherPresent, String other) {
-      return otherPresent ? other : mine;
-    }
-
-    @Override
-    public ByteString visitByteString(
-        boolean minePresent, ByteString mine, boolean otherPresent, ByteString other) {
-      return otherPresent ? other : mine;
-    }
-
-    @Override
-    public Object visitOneofBoolean(boolean minePresent, Object mine, Object other) {
-      return other;
-    }
-
-    @Override
-    public Object visitOneofInt(boolean minePresent, Object mine, Object other) {
-      return other;
-    }
-
-    @Override
-    public Object visitOneofDouble(boolean minePresent, Object mine, Object other) {
-      return other;
-    }
-
-    @Override
-    public Object visitOneofFloat(boolean minePresent, Object mine, Object other) {
-      return other;
-    }
-
-    @Override
-    public Object visitOneofLong(boolean minePresent, Object mine, Object other) {
-      return other;
-    }
-
-    @Override
-    public Object visitOneofString(boolean minePresent, Object mine, Object other) {
-      return other;
-    }
-
-    @Override
-    public Object visitOneofByteString(boolean minePresent, Object mine, Object other) {
-      return other;
-    }
-
-    @Override
-    public Object visitOneofLazyMessage(boolean minePresent, Object mine, Object other) {
-      LazyFieldLite lazy = minePresent ? (LazyFieldLite) mine : new LazyFieldLite();
-      lazy.merge((LazyFieldLite) other);
-      return lazy;
-    }
-
-    @Override
-    public Object visitOneofMessage(boolean minePresent, Object mine, Object other) {
-      if (minePresent) {
-        return visitMessage((MessageLite) mine, (MessageLite) other);
-      }
-      return other;
-    }
-
-    @Override
-    public void visitOneofNotSet(boolean minePresent) {
-      return;
-    }
-
-    @SuppressWarnings("unchecked") // Guaranteed by runtime.
-    @Override
-    public <T extends MessageLite> T visitMessage(T mine, T other) {
-      if (mine != null && other != null) {
-        return (T) mine.toBuilder().mergeFrom(other).build();
-      }
-
-      return mine != null ? mine : other;
-    }
-
-    @Override
-    public LazyFieldLite visitLazyMessage(LazyFieldLite mine, LazyFieldLite other) {
-      if (other != null) {
-        if (mine == null) {
-          mine = new LazyFieldLite();
-        }
-        mine.merge(other);
-      }
-      return mine;
-    }
-
-    @Override
-    public <T> ProtobufList<T> visitList(ProtobufList<T> mine, ProtobufList<T> other) {
-      int size = mine.size();
-      int otherSize = other.size();
-      if (size > 0 && otherSize > 0) {
-        if (!mine.isModifiable()) {
-          mine = mine.mutableCopyWithCapacity(size + otherSize);
-        }
-        mine.addAll(other);
-      }
-
-      return size > 0 ? mine : other;
-    }
-
-    @Override
-    public BooleanList visitBooleanList(BooleanList mine, BooleanList other) {
-      int size = mine.size();
-      int otherSize = other.size();
-      if (size > 0 && otherSize > 0) {
-        if (!mine.isModifiable()) {
-          mine = mine.mutableCopyWithCapacity(size + otherSize);
-        }
-        mine.addAll(other);
-      }
-
-      return size > 0 ? mine : other;
-    }
-
-    @Override
-    public IntList visitIntList(IntList mine, IntList other) {
-      int size = mine.size();
-      int otherSize = other.size();
-      if (size > 0 && otherSize > 0) {
-        if (!mine.isModifiable()) {
-          mine = mine.mutableCopyWithCapacity(size + otherSize);
-        }
-        mine.addAll(other);
-      }
-
-      return size > 0 ? mine : other;
-    }
-
-    @Override
-    public DoubleList visitDoubleList(DoubleList mine, DoubleList other) {
-      int size = mine.size();
-      int otherSize = other.size();
-      if (size > 0 && otherSize > 0) {
-        if (!mine.isModifiable()) {
-          mine = mine.mutableCopyWithCapacity(size + otherSize);
-        }
-        mine.addAll(other);
-      }
-
-      return size > 0 ? mine : other;
-    }
-
-    @Override
-    public FloatList visitFloatList(FloatList mine, FloatList other) {
-      int size = mine.size();
-      int otherSize = other.size();
-      if (size > 0 && otherSize > 0) {
-        if (!mine.isModifiable()) {
-          mine = mine.mutableCopyWithCapacity(size + otherSize);
-        }
-        mine.addAll(other);
-      }
-
-      return size > 0 ? mine : other;
-    }
-
-    @Override
-    public LongList visitLongList(LongList mine, LongList other) {
-      int size = mine.size();
-      int otherSize = other.size();
-      if (size > 0 && otherSize > 0) {
-        if (!mine.isModifiable()) {
-          mine = mine.mutableCopyWithCapacity(size + otherSize);
-        }
-        mine.addAll(other);
-      }
-
-      return size > 0 ? mine : other;
-    }
-
-    @Override
-    public FieldSet<ExtensionDescriptor> visitExtensions(
-        FieldSet<ExtensionDescriptor> mine,
-        FieldSet<ExtensionDescriptor> other) {
-      if (mine.isImmutable()) {
-        mine = mine.clone();
-      }
-      mine.mergeFrom(other);
-      return mine;
-    }
-
-    @Override
-    public UnknownFieldSetLite visitUnknownFields(
-        UnknownFieldSetLite mine,
-        UnknownFieldSetLite other) {
-      return other == UnknownFieldSetLite.getDefaultInstance()
-          ? mine : UnknownFieldSetLite.mutableCopyOf(mine, other);
-    }
-
-    @Override
-    public <K, V> MapFieldLite<K, V> visitMap(MapFieldLite<K, V> mine, MapFieldLite<K, V> other) {
-      if (!other.isEmpty()) {
-        if (!mine.isMutable()) {
-          mine = mine.mutableCopy();
-        }
-        mine.mergeFrom(other);
-      }
-      return mine;
-    }
-  }
-}


[30/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumValue.java
deleted file mode 100644
index a0a9d49..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumValue.java
+++ /dev/null
@@ -1,1044 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Enum value definition.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.EnumValue}
- */
-public  final class EnumValue extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.EnumValue)
-    EnumValueOrBuilder {
-  // Use EnumValue.newBuilder() to construct.
-  private EnumValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private EnumValue() {
-    name_ = "";
-    number_ = 0;
-    options_ = java.util.Collections.emptyList();
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private EnumValue(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            name_ = s;
-            break;
-          }
-          case 16: {
-
-            number_ = input.readInt32();
-            break;
-          }
-          case 26: {
-            if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-              options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>();
-              mutable_bitField0_ |= 0x00000004;
-            }
-            options_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.parser(), extensionRegistry));
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-        options_ = java.util.Collections.unmodifiableList(options_);
-      }
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_EnumValue_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_EnumValue_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder.class);
-  }
-
-  private int bitField0_;
-  public static final int NAME_FIELD_NUMBER = 1;
-  private volatile java.lang.Object name_;
-  /**
-   * <pre>
-   * Enum value name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public java.lang.String getName() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      name_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * Enum value name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      name_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int NUMBER_FIELD_NUMBER = 2;
-  private int number_;
-  /**
-   * <pre>
-   * Enum value number.
-   * </pre>
-   *
-   * <code>int32 number = 2;</code>
-   */
-  public int getNumber() {
-    return number_;
-  }
-
-  public static final int OPTIONS_FIELD_NUMBER = 3;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_;
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public int getOptionsCount() {
-    return options_.size();
-  }
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-    return options_.get(index);
-  }
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index) {
-    return options_.get(index);
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
-    }
-    if (number_ != 0) {
-      output.writeInt32(2, number_);
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      output.writeMessage(3, options_.get(i));
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
-    }
-    if (number_ != 0) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt32Size(2, number_);
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(3, options_.get(i));
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue) obj;
-
-    boolean result = true;
-    result = result && getName()
-        .equals(other.getName());
-    result = result && (getNumber()
-        == other.getNumber());
-    result = result && getOptionsList()
-        .equals(other.getOptionsList());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getName().hashCode();
-    hash = (37 * hash) + NUMBER_FIELD_NUMBER;
-    hash = (53 * hash) + getNumber();
-    if (getOptionsCount() > 0) {
-      hash = (37 * hash) + OPTIONS_FIELD_NUMBER;
-      hash = (53 * hash) + getOptionsList().hashCode();
-    }
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Enum value definition.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.EnumValue}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.EnumValue)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_EnumValue_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_EnumValue_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-        getOptionsFieldBuilder();
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      name_ = "";
-
-      number_ = 0;
-
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000004);
-      } else {
-        optionsBuilder_.clear();
-      }
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_EnumValue_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue(this);
-      int from_bitField0_ = bitField0_;
-      int to_bitField0_ = 0;
-      result.name_ = name_;
-      result.number_ = number_;
-      if (optionsBuilder_ == null) {
-        if (((bitField0_ & 0x00000004) == 0x00000004)) {
-          options_ = java.util.Collections.unmodifiableList(options_);
-          bitField0_ = (bitField0_ & ~0x00000004);
-        }
-        result.options_ = options_;
-      } else {
-        result.options_ = optionsBuilder_.build();
-      }
-      result.bitField0_ = to_bitField0_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.getDefaultInstance()) return this;
-      if (!other.getName().isEmpty()) {
-        name_ = other.name_;
-        onChanged();
-      }
-      if (other.getNumber() != 0) {
-        setNumber(other.getNumber());
-      }
-      if (optionsBuilder_ == null) {
-        if (!other.options_.isEmpty()) {
-          if (options_.isEmpty()) {
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000004);
-          } else {
-            ensureOptionsIsMutable();
-            options_.addAll(other.options_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.options_.isEmpty()) {
-          if (optionsBuilder_.isEmpty()) {
-            optionsBuilder_.dispose();
-            optionsBuilder_ = null;
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000004);
-            optionsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getOptionsFieldBuilder() : null;
-          } else {
-            optionsBuilder_.addAllMessages(other.options_);
-          }
-        }
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private java.lang.Object name_ = "";
-    /**
-     * <pre>
-     * Enum value name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        name_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * Enum value name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * Enum value name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      name_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder clearName() {
-      
-      name_ = getDefaultInstance().getName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      name_ = value;
-      onChanged();
-      return this;
-    }
-
-    private int number_ ;
-    /**
-     * <pre>
-     * Enum value number.
-     * </pre>
-     *
-     * <code>int32 number = 2;</code>
-     */
-    public int getNumber() {
-      return number_;
-    }
-    /**
-     * <pre>
-     * Enum value number.
-     * </pre>
-     *
-     * <code>int32 number = 2;</code>
-     */
-    public Builder setNumber(int value) {
-      
-      number_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value number.
-     * </pre>
-     *
-     * <code>int32 number = 2;</code>
-     */
-    public Builder clearNumber() {
-      
-      number_ = 0;
-      onChanged();
-      return this;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_ =
-      java.util.Collections.emptyList();
-    private void ensureOptionsIsMutable() {
-      if (!((bitField0_ & 0x00000004) == 0x00000004)) {
-        options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>(options_);
-        bitField0_ |= 0x00000004;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> optionsBuilder_;
-
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-      if (optionsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(options_);
-      } else {
-        return optionsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public int getOptionsCount() {
-      if (optionsBuilder_ == null) {
-        return options_.size();
-      } else {
-        return optionsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);
-      } else {
-        return optionsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.set(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addAllOptions(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> values) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, options_);
-        onChanged();
-      } else {
-        optionsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder clearOptions() {
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000004);
-        onChanged();
-      } else {
-        optionsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder removeOptions(int index) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.remove(index);
-        onChanged();
-      } else {
-        optionsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder getOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-        int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);  } else {
-        return optionsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-         getOptionsOrBuilderList() {
-      if (optionsBuilder_ != null) {
-        return optionsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(options_);
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder() {
-      return getOptionsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder> 
-         getOptionsBuilderList() {
-      return getOptionsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-        getOptionsFieldBuilder() {
-      if (optionsBuilder_ == null) {
-        optionsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder>(
-                options_,
-                ((bitField0_ & 0x00000004) == 0x00000004),
-                getParentForChildren(),
-                isClean());
-        options_ = null;
-      }
-      return optionsBuilder_;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.EnumValue)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.EnumValue)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EnumValue>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<EnumValue>() {
-    public EnumValue parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new EnumValue(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EnumValue> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<EnumValue> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumValueOrBuilder.java
deleted file mode 100644
index 22523d7..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumValueOrBuilder.java
+++ /dev/null
@@ -1,80 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface EnumValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.EnumValue)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * Enum value name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  java.lang.String getName();
-  /**
-   * <pre>
-   * Enum value name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes();
-
-  /**
-   * <pre>
-   * Enum value number.
-   * </pre>
-   *
-   * <code>int32 number = 2;</code>
-   */
-  int getNumber();
-
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> 
-      getOptionsList();
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index);
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  int getOptionsCount();
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList();
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExperimentalApi.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExperimentalApi.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExperimentalApi.java
deleted file mode 100644
index 3bfee9d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExperimentalApi.java
+++ /dev/null
@@ -1,66 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.lang.annotation.Documented;
-import java.lang.annotation.ElementType;
-import java.lang.annotation.Retention;
-import java.lang.annotation.RetentionPolicy;
-import java.lang.annotation.Target;
-
-/**
- * Indicates a public API that can change at any time, and has no guarantee of API stability and
- * backward-compatibility.
- *
- * <p>Usage guidelines:
- * <ol>
- * <li>This annotation is used only on public API. Internal interfaces should not use it.</li>
- * <li>This annotation should only be added to new APIs. Adding it to an existing API is
- * considered API-breaking.</li>
- * <li>Removing this annotation from an API gives it stable status.</li>
- * </ol>
- */
-@Retention(RetentionPolicy.SOURCE)
-@Target({
-    ElementType.ANNOTATION_TYPE,
-    ElementType.CONSTRUCTOR,
-    ElementType.FIELD,
-    ElementType.METHOD,
-    ElementType.PACKAGE,
-    ElementType.TYPE})
-@Documented
-public @interface ExperimentalApi {
-  /**
-   * Context information such as links to discussion thread, tracking issue etc.
-   */
-  String value() default "";
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Extension.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Extension.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Extension.java
deleted file mode 100644
index e5aa7d3..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Extension.java
+++ /dev/null
@@ -1,86 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Interface that generated extensions implement.
- *
- * @author liujisi@google.com (Jisi Liu)
- */
-public abstract class Extension<ContainingType extends MessageLite, Type>
-    extends ExtensionLite<ContainingType, Type> {
-
-  /** Returns the descriptor of the extension. */
-  public abstract Descriptors.FieldDescriptor getDescriptor();
-
-  /** Returns whether or not this extension is a Lite Extension. */
-  @Override
-  final boolean isLite() {
-    return false;
-  }
-
-  // All the methods below are extension implementation details.
-
-  /**
-   * The API type that the extension is used for.
-   */
-  protected enum ExtensionType {
-    IMMUTABLE,
-    MUTABLE,
-    PROTO1,
-  }
-
-  protected ExtensionType getExtensionType() {
-    // TODO(liujisi): make this abstract after we fix proto1.
-    return ExtensionType.IMMUTABLE;
-  }
-
-  /**
-   * Type of a message extension.
-   */
-  public enum MessageType {
-    PROTO1,
-    PROTO2,
-  }
-  
-  /**
-   * If the extension is a message extension (i.e., getLiteType() == MESSAGE),
-   * returns the type of the message, otherwise undefined.
-   */
-  public MessageType getMessageType() {
-    return MessageType.PROTO2;
-  }
-
-  protected abstract Object fromReflectionType(Object value);
-  protected abstract Object singularFromReflectionType(Object value);
-  protected abstract Object toReflectionType(Object value);
-  protected abstract Object singularToReflectionType(Object value);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionLite.java
deleted file mode 100644
index cb47243..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionLite.java
+++ /dev/null
@@ -1,63 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Lite interface that generated extensions implement.
- * <p>
- * Methods are for use by generated code only. You can hold a reference to
- * extensions using this type name.
- */
-public abstract class ExtensionLite<ContainingType extends MessageLite, Type> {
-
-  /** Returns the field number of the extension. */
-  public abstract int getNumber();
-
-  /** Returns the type of the field. */
-  public abstract WireFormat.FieldType getLiteType();
-
-  /** Returns whether it is a repeated field. */
-  public abstract boolean isRepeated();
-
-  /** Returns the default value of the extension field. */
-  public abstract Type getDefaultValue();
-
-  /**
-   * Returns the default instance of the extension field, if it's a message
-   * extension.
-   */
-  public abstract MessageLite getMessageDefaultInstance();
-  
-  /** Returns whether or not this extension is a Lite Extension. */
-  boolean isLite() {
-    return true;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistry.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistry.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistry.java
deleted file mode 100644
index c15eafd..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistry.java
+++ /dev/null
@@ -1,396 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-import java.util.Collection;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.HashSet;
-import java.util.Map;
-import java.util.Set;
-
-/**
- * A table of known extensions, searchable by name or field number.  When
- * parsing a protocol message that might have extensions, you must provide
- * an {@code ExtensionRegistry} in which you have registered any extensions
- * that you want to be able to parse.  Otherwise, those extensions will just
- * be treated like unknown fields.
- *
- * <p>For example, if you had the {@code .proto} file:
- *
- * <pre>
- * option java_class = "MyProto";
- *
- * message Foo {
- *   extensions 1000 to max;
- * }
- *
- * extend Foo {
- *   optional int32 bar;
- * }
- * </pre>
- *
- * Then you might write code like:
- *
- * <pre>
- * ExtensionRegistry registry = ExtensionRegistry.newInstance();
- * registry.add(MyProto.bar);
- * MyProto.Foo message = MyProto.Foo.parseFrom(input, registry);
- * </pre>
- *
- * <p>Background:
- *
- * <p>You might wonder why this is necessary.  Two alternatives might come to
- * mind.  First, you might imagine a system where generated extensions are
- * automatically registered when their containing classes are loaded.  This
- * is a popular technique, but is bad design; among other things, it creates a
- * situation where behavior can change depending on what classes happen to be
- * loaded.  It also introduces a security vulnerability, because an
- * unprivileged class could cause its code to be called unexpectedly from a
- * privileged class by registering itself as an extension of the right type.
- *
- * <p>Another option you might consider is lazy parsing: do not parse an
- * extension until it is first requested, at which point the caller must
- * provide a type to use.  This introduces a different set of problems.  First,
- * it would require a mutex lock any time an extension was accessed, which
- * would be slow.  Second, corrupt data would not be detected until first
- * access, at which point it would be much harder to deal with it.  Third, it
- * could violate the expectation that message objects are immutable, since the
- * type provided could be any arbitrary message class.  An unprivileged user
- * could take advantage of this to inject a mutable object into a message
- * belonging to privileged code and create mischief.
- *
- * @author kenton@google.com Kenton Varda
- */
-public class ExtensionRegistry extends ExtensionRegistryLite {
-  /** Construct a new, empty instance. */
-  public static ExtensionRegistry newInstance() {
-    return new ExtensionRegistry();
-  }
-
-  /** Get the unmodifiable singleton empty instance. */
-  public static ExtensionRegistry getEmptyRegistry() {
-    return EMPTY_REGISTRY;
-  }
-
-
-  /** Returns an unmodifiable view of the registry. */
-  @Override
-  public ExtensionRegistry getUnmodifiable() {
-    return new ExtensionRegistry(this);
-  }
-
-  /** A (Descriptor, Message) pair, returned by lookup methods. */
-  public static final class ExtensionInfo {
-    /** The extension's descriptor. */
-    public final FieldDescriptor descriptor;
-
-    /**
-     * A default instance of the extension's type, if it has a message type.
-     * Otherwise, {@code null}.
-     */
-    public final Message defaultInstance;
-
-    private ExtensionInfo(final FieldDescriptor descriptor) {
-      this.descriptor = descriptor;
-      defaultInstance = null;
-    }
-    private ExtensionInfo(final FieldDescriptor descriptor,
-                          final Message defaultInstance) {
-      this.descriptor = descriptor;
-      this.defaultInstance = defaultInstance;
-    }
-  }
-
-  /**
-   * Deprecated. Use {@link #findImmutableExtensionByName(String)} instead.
-   */
-  public ExtensionInfo findExtensionByName(final String fullName) {
-    return findImmutableExtensionByName(fullName);
-  }
-
-  /**
-   * Find an extension for immutable APIs by fully-qualified field name,
-   * in the proto namespace. i.e. {@code result.descriptor.fullName()} will
-   * match {@code fullName} if a match is found.
-   *
-   * @return Information about the extension if found, or {@code null}
-   *         otherwise.
-   */
-  public ExtensionInfo findImmutableExtensionByName(final String fullName) {
-    return immutableExtensionsByName.get(fullName);
-  }
-
-  /**
-   * Find an extension for mutable APIs by fully-qualified field name,
-   * in the proto namespace. i.e. {@code result.descriptor.fullName()} will
-   * match {@code fullName} if a match is found.
-   *
-   * @return Information about the extension if found, or {@code null}
-   *         otherwise.
-   */
-  public ExtensionInfo findMutableExtensionByName(final String fullName) {
-    return mutableExtensionsByName.get(fullName);
-  }
-
-  /**
-   * Deprecated. Use {@link #findImmutableExtensionByNumber(
-   * Descriptors.Descriptor, int)}
-   */
-  public ExtensionInfo findExtensionByNumber(
-      final Descriptor containingType, final int fieldNumber) {
-    return findImmutableExtensionByNumber(containingType, fieldNumber);
-  }
-
-  /**
-   * Find an extension by containing type and field number for immutable APIs.
-   *
-   * @return Information about the extension if found, or {@code null}
-   *         otherwise.
-   */
-  public ExtensionInfo findImmutableExtensionByNumber(
-      final Descriptor containingType, final int fieldNumber) {
-    return immutableExtensionsByNumber.get(
-      new DescriptorIntPair(containingType, fieldNumber));
-  }
-
-  /**
-   * Find an extension by containing type and field number for mutable APIs.
-   *
-   * @return Information about the extension if found, or {@code null}
-   *         otherwise.
-   */
-  public ExtensionInfo findMutableExtensionByNumber(
-      final Descriptor containingType, final int fieldNumber) {
-    return mutableExtensionsByNumber.get(
-        new DescriptorIntPair(containingType, fieldNumber));
-  }
-
-  /**
-   * Find all extensions for mutable APIs by fully-qualified name of
-   * extended class. Note that this method is more computationally expensive
-   * than getting a single extension by name or number.
-   *
-   * @return Information about the extensions found, or {@code null} if there
-   *     are none.
-   */
-  public Set<ExtensionInfo> getAllMutableExtensionsByExtendedType(final String fullName) {
-    HashSet<ExtensionInfo> extensions = new HashSet<ExtensionInfo>();
-    for (DescriptorIntPair pair : mutableExtensionsByNumber.keySet()) {
-      if (pair.descriptor.getFullName().equals(fullName)) {
-        extensions.add(mutableExtensionsByNumber.get(pair));
-      }
-    }
-    return extensions;
-  }
-
-  /**
-   * Find all extensions for immutable APIs by fully-qualified name of
-   * extended class. Note that this method is more computationally expensive
-   * than getting a single extension by name or number.
-   *
-   * @return Information about the extensions found, or {@code null} if there
-   *     are none.
-   */
-  public Set<ExtensionInfo> getAllImmutableExtensionsByExtendedType(final String fullName) {
-    HashSet<ExtensionInfo> extensions = new HashSet<ExtensionInfo>();
-    for (DescriptorIntPair pair : immutableExtensionsByNumber.keySet()) {
-      if (pair.descriptor.getFullName().equals(fullName)) {
-        extensions.add(immutableExtensionsByNumber.get(pair));
-      }
-    }
-    return extensions;
-  }
-
-  /** Add an extension from a generated file to the registry. */
-  public void add(final Extension<?, ?> extension) {
-    if (extension.getExtensionType() != Extension.ExtensionType.IMMUTABLE &&
-        extension.getExtensionType() != Extension.ExtensionType.MUTABLE) {
-      // do not support other extension types. ignore
-      return;
-    }
-    add(newExtensionInfo(extension), extension.getExtensionType());
-  }
-
-  /** Add an extension from a generated file to the registry. */
-  public void add(final GeneratedMessage.GeneratedExtension<?, ?> extension) {
-    add((Extension<?, ?>) extension);
-  }
-
-  static ExtensionInfo newExtensionInfo(final Extension<?, ?> extension) {
-    if (extension.getDescriptor().getJavaType() ==
-        FieldDescriptor.JavaType.MESSAGE) {
-      if (extension.getMessageDefaultInstance() == null) {
-        throw new IllegalStateException(
-            "Registered message-type extension had null default instance: " +
-                extension.getDescriptor().getFullName());
-      }
-      return new ExtensionInfo(extension.getDescriptor(),
-          (Message) extension.getMessageDefaultInstance());
-    } else {
-      return new ExtensionInfo(extension.getDescriptor(), null);
-    }
-  }
-
-  /** Add a non-message-type extension to the registry by descriptor. */
-  public void add(final FieldDescriptor type) {
-    if (type.getJavaType() == FieldDescriptor.JavaType.MESSAGE) {
-      throw new IllegalArgumentException(
-        "ExtensionRegistry.add() must be provided a default instance when " +
-        "adding an embedded message extension.");
-    }
-    ExtensionInfo info = new ExtensionInfo(type, null);
-    add(info, Extension.ExtensionType.IMMUTABLE);
-    add(info, Extension.ExtensionType.MUTABLE);
-  }
-
-  /** Add a message-type extension to the registry by descriptor. */
-  public void add(final FieldDescriptor type, final Message defaultInstance) {
-    if (type.getJavaType() != FieldDescriptor.JavaType.MESSAGE) {
-      throw new IllegalArgumentException(
-        "ExtensionRegistry.add() provided a default instance for a " +
-        "non-message extension.");
-    }
-      add(new ExtensionInfo(type, defaultInstance),
-          Extension.ExtensionType.IMMUTABLE);
-  }
-
-  // =================================================================
-  // Private stuff.
-
-  private ExtensionRegistry() {
-    this.immutableExtensionsByName = new HashMap<String, ExtensionInfo>();
-    this.mutableExtensionsByName = new HashMap<String, ExtensionInfo>();
-    this.immutableExtensionsByNumber =
-        new HashMap<DescriptorIntPair, ExtensionInfo>();
-    this.mutableExtensionsByNumber =
-        new HashMap<DescriptorIntPair, ExtensionInfo>();
-  }
-
-  private ExtensionRegistry(ExtensionRegistry other) {
-    super(other);
-    this.immutableExtensionsByName =
-        Collections.unmodifiableMap(other.immutableExtensionsByName);
-    this.mutableExtensionsByName =
-        Collections.unmodifiableMap(other.mutableExtensionsByName);
-    this.immutableExtensionsByNumber =
-        Collections.unmodifiableMap(other.immutableExtensionsByNumber);
-    this.mutableExtensionsByNumber =
-            Collections.unmodifiableMap(other.mutableExtensionsByNumber);
-  }
-
-  private final Map<String, ExtensionInfo> immutableExtensionsByName;
-  private final Map<String, ExtensionInfo> mutableExtensionsByName;
-  private final Map<DescriptorIntPair, ExtensionInfo> immutableExtensionsByNumber;
-  private final Map<DescriptorIntPair, ExtensionInfo> mutableExtensionsByNumber;
-
-  ExtensionRegistry(boolean empty) {
-    super(EMPTY_REGISTRY_LITE);
-    this.immutableExtensionsByName =
-        Collections.<String, ExtensionInfo>emptyMap();
-    this.mutableExtensionsByName =
-        Collections.<String, ExtensionInfo>emptyMap();
-    this.immutableExtensionsByNumber =
-        Collections.<DescriptorIntPair, ExtensionInfo>emptyMap();
-    this.mutableExtensionsByNumber =
-            Collections.<DescriptorIntPair, ExtensionInfo>emptyMap();
-  }
-  static final ExtensionRegistry EMPTY_REGISTRY = new ExtensionRegistry(true);
-
-  private void add(
-      final ExtensionInfo extension,
-      final Extension.ExtensionType extensionType) {
-    if (!extension.descriptor.isExtension()) {
-      throw new IllegalArgumentException(
-          "ExtensionRegistry.add() was given a FieldDescriptor for a regular " +
-              "(non-extension) field.");
-    }
-
-    Map<String, ExtensionInfo> extensionsByName;
-    Map<DescriptorIntPair, ExtensionInfo> extensionsByNumber;
-    switch (extensionType) {
-      case IMMUTABLE:
-        extensionsByName = immutableExtensionsByName;
-        extensionsByNumber = immutableExtensionsByNumber;
-        break;
-      case MUTABLE:
-        extensionsByName = mutableExtensionsByName;
-        extensionsByNumber = mutableExtensionsByNumber;
-        break;
-      default:
-        // Ignore the unknown supported type.
-        return;
-    }
-
-    extensionsByName.put(extension.descriptor.getFullName(), extension);
-    extensionsByNumber.put(
-      new DescriptorIntPair(extension.descriptor.getContainingType(),
-                            extension.descriptor.getNumber()),
-      extension);
-
-    final FieldDescriptor field = extension.descriptor;
-    if (field.getContainingType().getOptions().getMessageSetWireFormat() &&
-        field.getType() == FieldDescriptor.Type.MESSAGE &&
-        field.isOptional() &&
-        field.getExtensionScope() == field.getMessageType()) {
-      // This is an extension of a MessageSet type defined within the extension
-      // type's own scope.  For backwards-compatibility, allow it to be looked
-      // up by type name.
-      extensionsByName.put(field.getMessageType().getFullName(), extension);
-    }
-  }
-
-  /** A (GenericDescriptor, int) pair, used as a map key. */
-  private static final class DescriptorIntPair {
-    private final Descriptor descriptor;
-    private final int number;
-
-    DescriptorIntPair(final Descriptor descriptor, final int number) {
-      this.descriptor = descriptor;
-      this.number = number;
-    }
-
-    @Override
-    public int hashCode() {
-      return descriptor.hashCode() * ((1 << 16) - 1) + number;
-    }
-    @Override
-    public boolean equals(final Object obj) {
-      if (!(obj instanceof DescriptorIntPair)) {
-        return false;
-      }
-      final DescriptorIntPair other = (DescriptorIntPair)obj;
-      return descriptor == other.descriptor && number == other.number;
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistryFactory.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistryFactory.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistryFactory.java
deleted file mode 100644
index 85e6668..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistryFactory.java
+++ /dev/null
@@ -1,95 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite.EMPTY_REGISTRY_LITE;
-
-/**
- * A factory object to create instances of {@link ExtensionRegistryLite}.
- * 
- * <p>
- * This factory detects (via reflection) if the full (non-Lite) protocol buffer libraries
- * are available, and if so, the instances returned are actually {@link ExtensionRegistry}.
- */
-final class ExtensionRegistryFactory {
-
-  static final String FULL_REGISTRY_CLASS_NAME = "org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry";
-
-  /* Visible for Testing
-     @Nullable */
-  static final Class<?> EXTENSION_REGISTRY_CLASS = reflectExtensionRegistry();
-
-  /* @Nullable */
-  static Class<?> reflectExtensionRegistry() {
-    try {
-      return Class.forName(FULL_REGISTRY_CLASS_NAME);
-    } catch (ClassNotFoundException e) {
-      // The exception allocation is potentially expensive on Android (where it can be triggered
-      // many times at start up). Is there a way to ameliorate this?
-      return null;
-    }
-  }
-
-  /** Construct a new, empty instance. */
-  public static ExtensionRegistryLite create() {
-    if (EXTENSION_REGISTRY_CLASS != null) {
-      try {
-        return invokeSubclassFactory("newInstance");
-      } catch (Exception e) {
-        // return a Lite registry.
-      }
-    }
-    return new ExtensionRegistryLite();
-  }
-
-  /** Get the unmodifiable singleton empty instance. */
-  public static ExtensionRegistryLite createEmpty() {
-    if (EXTENSION_REGISTRY_CLASS != null) {
-      try {
-        return invokeSubclassFactory("getEmptyRegistry");
-      } catch (Exception e) {
-        // return a Lite registry.
-      }
-    }
-    return EMPTY_REGISTRY_LITE;
-  }
-
-  static boolean isFullRegistry(ExtensionRegistryLite registry) {
-    return EXTENSION_REGISTRY_CLASS != null
-        && EXTENSION_REGISTRY_CLASS.isAssignableFrom(registry.getClass());
-  }
-
-  private static final ExtensionRegistryLite invokeSubclassFactory(String methodName)
-      throws Exception {
-    return (ExtensionRegistryLite) EXTENSION_REGISTRY_CLASS
-        .getMethod(methodName).invoke(null);
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistryLite.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistryLite.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistryLite.java
deleted file mode 100644
index 9bf452a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ExtensionRegistryLite.java
+++ /dev/null
@@ -1,227 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.Map;
-
-/**
- * Equivalent to {@link ExtensionRegistry} but supports only "lite" types.
- * <p>
- * If all of your types are lite types, then you only need to use
- * {@code ExtensionRegistryLite}.  Similarly, if all your types are regular
- * types, then you only need {@link ExtensionRegistry}.  Typically it does not
- * make sense to mix the two, since if you have any regular types in your
- * program, you then require the full runtime and lose all the benefits of
- * the lite runtime, so you might as well make all your types be regular types.
- * However, in some cases (e.g. when depending on multiple third-party libraries
- * where one uses lite types and one uses regular), you may find yourself
- * wanting to mix the two.  In this case things get more complicated.
- * <p>
- * There are three factors to consider:  Whether the type being extended is
- * lite, whether the embedded type (in the case of a message-typed extension)
- * is lite, and whether the extension itself is lite.  Since all three are
- * declared in different files, they could all be different.  Here are all
- * the combinations and which type of registry to use:
- * <pre>
- *   Extended type     Inner type    Extension         Use registry
- *   =======================================================================
- *   lite              lite          lite              ExtensionRegistryLite
- *   lite              regular       lite              ExtensionRegistry
- *   regular           regular       regular           ExtensionRegistry
- *   all other combinations                            not supported
- * </pre>
- * <p>
- * Note that just as regular types are not allowed to contain lite-type fields,
- * they are also not allowed to contain lite-type extensions.  This is because
- * regular types must be fully accessible via reflection, which in turn means
- * that all the inner messages must also support reflection.  On the other hand,
- * since regular types implement the entire lite interface, there is no problem
- * with embedding regular types inside lite types.
- *
- * @author kenton@google.com Kenton Varda
- */
-public class ExtensionRegistryLite {
-
-  // Set true to enable lazy parsing feature for MessageSet.
-  //
-  // TODO(xiangl): Now we use a global flag to control whether enable lazy
-  // parsing feature for MessageSet, which may be too crude for some
-  // applications. Need to support this feature on smaller granularity.
-  private static volatile boolean eagerlyParseMessageSets = false;
-
-  // Visible for testing.
-  static final String EXTENSION_CLASS_NAME = "org.apache.hadoop.hbase.shaded.com.google.protobuf.Extension";
-
-  /* @Nullable */
-  static Class<?> resolveExtensionClass() {
-    try {
-      return Class.forName(EXTENSION_CLASS_NAME);
-    } catch (ClassNotFoundException e) {
-      // See comment in ExtensionRegistryFactory on the potential expense of this.
-      return null;
-    }
-  }
-
-  /* @Nullable */
-  private static final Class<?> extensionClass = resolveExtensionClass();
-
-  public static boolean isEagerlyParseMessageSets() {
-    return eagerlyParseMessageSets;
-  }
-
-  public static void setEagerlyParseMessageSets(boolean isEagerlyParse) {
-    eagerlyParseMessageSets = isEagerlyParse;
-  }
-
-  /**
-   * Construct a new, empty instance.
-   *
-   * <p>This may be an {@code ExtensionRegistry} if the full (non-Lite) proto libraries are
-   * available.
-   */
-  public static ExtensionRegistryLite newInstance() {
-    return ExtensionRegistryFactory.create();
-  }
-
-  /**
-   * Get the unmodifiable singleton empty instance of either ExtensionRegistryLite or
-   * {@code ExtensionRegistry} (if the full (non-Lite) proto libraries are available).
-   */
-  public static ExtensionRegistryLite getEmptyRegistry() {
-    return ExtensionRegistryFactory.createEmpty();
-  }
-
-
-  /** Returns an unmodifiable view of the registry. */
-  public ExtensionRegistryLite getUnmodifiable() {
-    return new ExtensionRegistryLite(this);
-  }
-
-  /**
-   * Find an extension by containing type and field number.
-   *
-   * @return Information about the extension if found, or {@code null}
-   *         otherwise.
-   */
-  @SuppressWarnings("unchecked")
-  public <ContainingType extends MessageLite>
-      GeneratedMessageLite.GeneratedExtension<ContainingType, ?>
-        findLiteExtensionByNumber(
-          final ContainingType containingTypeDefaultInstance,
-          final int fieldNumber) {
-    return (GeneratedMessageLite.GeneratedExtension<ContainingType, ?>)
-      extensionsByNumber.get(
-        new ObjectIntPair(containingTypeDefaultInstance, fieldNumber));
-  }
-
-  /** Add an extension from a lite generated file to the registry. */
-  public final void add(
-      final GeneratedMessageLite.GeneratedExtension<?, ?> extension) {
-    extensionsByNumber.put(
-      new ObjectIntPair(extension.getContainingTypeDefaultInstance(),
-                        extension.getNumber()),
-      extension);
-  }
-
-  /**
-   * Add an extension from a lite generated file to the registry only if it is
-   * a non-lite extension i.e. {@link GeneratedMessageLite.GeneratedExtension}. */
-  public final void add(ExtensionLite<?, ?> extension) {
-    if (GeneratedMessageLite.GeneratedExtension.class.isAssignableFrom(extension.getClass())) {
-      add((GeneratedMessageLite.GeneratedExtension<?, ?>) extension);
-    }
-    if (ExtensionRegistryFactory.isFullRegistry(this)) {
-      try {
-        this.getClass().getMethod("add", extensionClass).invoke(this, extension);
-      } catch (Exception e) {
-        throw new IllegalArgumentException(
-            String.format("Could not invoke ExtensionRegistry#add for %s", extension), e);
-      }
-    }
-  }
-
-  // =================================================================
-  // Private stuff.
-
-  // Constructors are package-private so that ExtensionRegistry can subclass
-  // this.
-
-  ExtensionRegistryLite() {
-    this.extensionsByNumber =
-        new HashMap<ObjectIntPair,
-                    GeneratedMessageLite.GeneratedExtension<?, ?>>();
-  }
-  static final ExtensionRegistryLite EMPTY_REGISTRY_LITE =
-      new ExtensionRegistryLite(true);
-
-  ExtensionRegistryLite(ExtensionRegistryLite other) {
-    if (other == EMPTY_REGISTRY_LITE) {
-      this.extensionsByNumber = Collections.emptyMap();
-    } else {
-      this.extensionsByNumber =
-        Collections.unmodifiableMap(other.extensionsByNumber);
-    }
-  }
-
-  private final Map<ObjectIntPair,
-                    GeneratedMessageLite.GeneratedExtension<?, ?>>
-      extensionsByNumber;
-
-  ExtensionRegistryLite(boolean empty) {
-    this.extensionsByNumber = Collections.emptyMap();
-  }
-
-  /** A (Object, int) pair, used as a map key. */
-  private static final class ObjectIntPair {
-    private final Object object;
-    private final int number;
-
-    ObjectIntPair(final Object object, final int number) {
-      this.object = object;
-      this.number = number;
-    }
-
-    @Override
-    public int hashCode() {
-      return System.identityHashCode(object) * ((1 << 16) - 1) + number;
-    }
-    @Override
-    public boolean equals(final Object obj) {
-      if (!(obj instanceof ObjectIntPair)) {
-        return false;
-      }
-      final ObjectIntPair other = (ObjectIntPair)obj;
-      return object == other.object && number == other.number;
-    }
-  }
-}


[38/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java
deleted file mode 100644
index b8b7030..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java
+++ /dev/null
@@ -1,258 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface ApiOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Api)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The fully qualified name of this api, including package name
-   * followed by the api's simple name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  java.lang.String getName();
-  /**
-   * <pre>
-   * The fully qualified name of this api, including package name
-   * followed by the api's simple name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes();
-
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method> 
-      getMethodsList();
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Method getMethods(int index);
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  int getMethodsCount();
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder> 
-      getMethodsOrBuilderList();
-  /**
-   * <pre>
-   * The methods of this api, in unspecified order.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Method methods = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder getMethodsOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> 
-      getOptionsList();
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index);
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  int getOptionsCount();
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList();
-  /**
-   * <pre>
-   * Any metadata attached to the API.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * A version string for this api. If specified, must have the form
-   * `major-version.minor-version`, as in `1.10`. If the minor version
-   * is omitted, it defaults to zero. If the entire version field is
-   * empty, the major version is derived from the package name, as
-   * outlined below. If the field is not empty, the version in the
-   * package name will be verified to be consistent with what is
-   * provided here.
-   * The versioning schema uses [semantic
-   * versioning](http://semver.org) where the major version number
-   * indicates a breaking change and the minor version an additive,
-   * non-breaking change. Both version numbers are signals to users
-   * what to expect from different versions, and should be carefully
-   * chosen based on the product plan.
-   * The major version is also reflected in the package name of the
-   * API, which must end in `v&lt;major-version&gt;`, as in
-   * `google.feature.v1`. For major versions 0 and 1, the suffix can
-   * be omitted. Zero major versions must only be used for
-   * experimental, none-GA apis.
-   * </pre>
-   *
-   * <code>string version = 4;</code>
-   */
-  java.lang.String getVersion();
-  /**
-   * <pre>
-   * A version string for this api. If specified, must have the form
-   * `major-version.minor-version`, as in `1.10`. If the minor version
-   * is omitted, it defaults to zero. If the entire version field is
-   * empty, the major version is derived from the package name, as
-   * outlined below. If the field is not empty, the version in the
-   * package name will be verified to be consistent with what is
-   * provided here.
-   * The versioning schema uses [semantic
-   * versioning](http://semver.org) where the major version number
-   * indicates a breaking change and the minor version an additive,
-   * non-breaking change. Both version numbers are signals to users
-   * what to expect from different versions, and should be carefully
-   * chosen based on the product plan.
-   * The major version is also reflected in the package name of the
-   * API, which must end in `v&lt;major-version&gt;`, as in
-   * `google.feature.v1`. For major versions 0 and 1, the suffix can
-   * be omitted. Zero major versions must only be used for
-   * experimental, none-GA apis.
-   * </pre>
-   *
-   * <code>string version = 4;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getVersionBytes();
-
-  /**
-   * <pre>
-   * Source context for the protocol buffer service represented by this
-   * message.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  boolean hasSourceContext();
-  /**
-   * <pre>
-   * Source context for the protocol buffer service represented by this
-   * message.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext();
-  /**
-   * <pre>
-   * Source context for the protocol buffer service represented by this
-   * message.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 5;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder();
-
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin> 
-      getMixinsList();
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin getMixins(int index);
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  int getMixinsCount();
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder> 
-      getMixinsOrBuilderList();
-  /**
-   * <pre>
-   * Included APIs. See [Mixin][].
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Mixin mixins = 6;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder getMixinsOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * The source syntax of the service.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 7;</code>
-   */
-  int getSyntaxValue();
-  /**
-   * <pre>
-   * The source syntax of the service.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 7;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java
deleted file mode 100644
index 817e07d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java
+++ /dev/null
@@ -1,98 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class ApiProto {
-  private ApiProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Api_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Api_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Method_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Method_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Mixin_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Mixin_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\031google/protobuf/api.proto\022\017google.prot" +
-      "obuf\032$google/protobuf/source_context.pro" +
-      "to\032\032google/protobuf/type.proto\"\201\002\n\003Api\022\014" +
-      "\n\004name\030\001 \001(\t\022(\n\007methods\030\002 \003(\0132\027.google.p" +
-      "rotobuf.Method\022(\n\007options\030\003 \003(\0132\027.google" +
-      ".protobuf.Option\022\017\n\007version\030\004 \001(\t\0226\n\016sou" +
-      "rce_context\030\005 \001(\0132\036.google.protobuf.Sour" +
-      "ceContext\022&\n\006mixins\030\006 \003(\0132\026.google.proto" +
-      "buf.Mixin\022\'\n\006syntax\030\007 \001(\0162\027.google.proto" +
-      "buf.Syntax\"\325\001\n\006Method\022\014\n\004name\030\001 \001(\t\022\030\n\020r",
-      "equest_type_url\030\002 \001(\t\022\031\n\021request_streami" +
-      "ng\030\003 \001(\010\022\031\n\021response_type_url\030\004 \001(\t\022\032\n\022r" +
-      "esponse_streaming\030\005 \001(\010\022(\n\007options\030\006 \003(\013" +
-      "2\027.google.protobuf.Option\022\'\n\006syntax\030\007 \001(" +
-      "\0162\027.google.protobuf.Syntax\"#\n\005Mixin\022\014\n\004n" +
-      "ame\030\001 \001(\t\022\014\n\004root\030\002 \001(\tBu\n\023com.google.pr" +
-      "otobufB\010ApiProtoP\001Z+google.golang.org/ge" +
-      "nproto/protobuf/api;api\242\002\003GPB\252\002\036Google.P" +
-      "rotobuf.WellKnownTypesb\006proto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.getDescriptor(),
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.getDescriptor(),
-        }, assigner);
-    internal_static_google_protobuf_Api_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_Api_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Api_descriptor,
-        new java.lang.String[] { "Name", "Methods", "Options", "Version", "SourceContext", "Mixins", "Syntax", });
-    internal_static_google_protobuf_Method_descriptor =
-      getDescriptor().getMessageTypes().get(1);
-    internal_static_google_protobuf_Method_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Method_descriptor,
-        new java.lang.String[] { "Name", "RequestTypeUrl", "RequestStreaming", "ResponseTypeUrl", "ResponseStreaming", "Options", "Syntax", });
-    internal_static_google_protobuf_Mixin_descriptor =
-      getDescriptor().getMessageTypes().get(2);
-    internal_static_google_protobuf_Mixin_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Mixin_descriptor,
-        new java.lang.String[] { "Name", "Root", });
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.getDescriptor();
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.getDescriptor();
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java
deleted file mode 100644
index bd2f2cd..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java
+++ /dev/null
@@ -1,51 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <p>Abstract interface for a blocking RPC channel.  {@code BlockingRpcChannel}
- * is the blocking equivalent to {@link RpcChannel}.
- *
- * @author kenton@google.com Kenton Varda
- * @author cpovirk@google.com Chris Povirk
- */
-public interface BlockingRpcChannel {
-  /**
-   * Call the given method of the remote service and blocks until it returns.
-   * {@code callBlockingMethod()} is the blocking equivalent to
-   * {@link RpcChannel#callMethod}.
-   */
-  Message callBlockingMethod(
-      Descriptors.MethodDescriptor method,
-      RpcController controller,
-      Message request,
-      Message responsePrototype) throws ServiceException;
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java
deleted file mode 100644
index f496f2d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java
+++ /dev/null
@@ -1,64 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Blocking equivalent to {@link Service}.
- *
- * @author kenton@google.com Kenton Varda
- * @author cpovirk@google.com Chris Povirk
- */
-public interface BlockingService {
-  /**
-   * Equivalent to {@link Service#getDescriptorForType}.
-   */
-  Descriptors.ServiceDescriptor getDescriptorForType();
-
-  /**
-   * Equivalent to {@link Service#callMethod}, except that
-   * {@code callBlockingMethod()} returns the result of the RPC or throws a
-   * {@link ServiceException} if there is a failure, rather than passing the
-   * information to a callback.
-   */
-  Message callBlockingMethod(Descriptors.MethodDescriptor method,
-                             RpcController controller,
-                             Message request) throws ServiceException;
-
-  /**
-   * Equivalent to {@link Service#getRequestPrototype}.
-   */
-  Message getRequestPrototype(Descriptors.MethodDescriptor method);
-
-  /**
-   * Equivalent to {@link Service#getResponsePrototype}.
-   */
-  Message getResponsePrototype(Descriptors.MethodDescriptor method);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java
deleted file mode 100644
index a2dbe4a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java
+++ /dev/null
@@ -1,452 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `bool`.
- * The JSON representation for `BoolValue` is JSON `true` and `false`.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.BoolValue}
- */
-public  final class BoolValue extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.BoolValue)
-    BoolValueOrBuilder {
-  // Use BoolValue.newBuilder() to construct.
-  private BoolValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private BoolValue() {
-    value_ = false;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private BoolValue(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-
-            value_ = input.readBool();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private boolean value_;
-  /**
-   * <pre>
-   * The bool value.
-   * </pre>
-   *
-   * <code>bool value = 1;</code>
-   */
-  public boolean getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (value_ != false) {
-      output.writeBool(1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (value_ != false) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeBoolSize(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) obj;
-
-    boolean result = true;
-    result = result && (getValue()
-        == other.getValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-        getValue());
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `bool`.
-   * The JSON representation for `BoolValue` is JSON `true` and `false`.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.BoolValue}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.BoolValue)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = false;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.getDefaultInstance()) return this;
-      if (other.getValue() != false) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private boolean value_ ;
-    /**
-     * <pre>
-     * The bool value.
-     * </pre>
-     *
-     * <code>bool value = 1;</code>
-     */
-    public boolean getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * The bool value.
-     * </pre>
-     *
-     * <code>bool value = 1;</code>
-     */
-    public Builder setValue(boolean value) {
-      
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The bool value.
-     * </pre>
-     *
-     * <code>bool value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = false;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.BoolValue)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.BoolValue)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<BoolValue>() {
-    public BoolValue parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new BoolValue(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java
deleted file mode 100644
index c1d3b5f..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface BoolValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.BoolValue)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The bool value.
-   * </pre>
-   *
-   * <code>bool value = 1;</code>
-   */
-  boolean getValue();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java
deleted file mode 100644
index 6bc9641..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java
+++ /dev/null
@@ -1,272 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.BooleanList;
-
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.RandomAccess;
-
-/**
- * An implementation of {@link BooleanList} on top of a primitive array.
- *
- * @author dweis@google.com (Daniel Weis)
- */
-final class BooleanArrayList
-    extends AbstractProtobufList<Boolean>
-    implements BooleanList, RandomAccess {
-
-  private static final BooleanArrayList EMPTY_LIST = new BooleanArrayList();
-  static {
-    EMPTY_LIST.makeImmutable();
-  }
-
-  public static BooleanArrayList emptyList() {
-    return EMPTY_LIST;
-  }
-
-  /**
-   * The backing store for the list.
-   */
-  private boolean[] array;
-
-  /**
-   * The size of the list distinct from the length of the array. That is, it is the number of
-   * elements set in the list.
-   */
-  private int size;
-
-  /**
-   * Constructs a new mutable {@code BooleanArrayList} with default capacity.
-   */
-  BooleanArrayList() {
-    this(new boolean[DEFAULT_CAPACITY], 0);
-  }
-
-  /**
-   * Constructs a new mutable {@code BooleanArrayList}
-   * containing the same elements as {@code other}.
-   */
-  private BooleanArrayList(boolean[] other, int size) {
-    array = other;
-    this.size = size;
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (this == o) {
-      return true;
-    }
-    if (!(o instanceof BooleanArrayList)) {
-      return super.equals(o);
-    }
-    BooleanArrayList other = (BooleanArrayList) o;
-    if (size != other.size) {
-      return false;
-    }
-
-    final boolean[] arr = other.array;
-    for (int i = 0; i < size; i++) {
-      if (array[i] != arr[i]) {
-        return false;
-      }
-    }
-
-    return true;
-  }
-
-  @Override
-  public int hashCode() {
-    int result = 1;
-    for (int i = 0; i < size; i++) {
-      result = (31 * result) + Internal.hashBoolean(array[i]);
-    }
-    return result;
-  }
-
-  @Override
-  public BooleanList mutableCopyWithCapacity(int capacity) {
-    if (capacity < size) {
-      throw new IllegalArgumentException();
-    }
-    return new BooleanArrayList(Arrays.copyOf(array, capacity), size);
-  }
-
-  @Override
-  public Boolean get(int index) {
-    return getBoolean(index);
-  }
-
-  @Override
-  public boolean getBoolean(int index) {
-    ensureIndexInRange(index);
-    return array[index];
-  }
-
-  @Override
-  public int size() {
-    return size;
-  }
-
-  @Override
-  public Boolean set(int index, Boolean element) {
-    return setBoolean(index, element);
-  }
-
-  @Override
-  public boolean setBoolean(int index, boolean element) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    boolean previousValue = array[index];
-    array[index] = element;
-    return previousValue;
-  }
-
-  @Override
-  public void add(int index, Boolean element) {
-    addBoolean(index, element);
-  }
-
-  /**
-   * Like {@link #add(Boolean)} but more efficient in that it doesn't box the element.
-   */
-  @Override
-  public void addBoolean(boolean element) {
-    addBoolean(size, element);
-  }
-
-  /**
-   * Like {@link #add(int, Boolean)} but more efficient in that it doesn't box the element.
-   */
-  private void addBoolean(int index, boolean element) {
-    ensureIsMutable();
-    if (index < 0 || index > size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-
-    if (size < array.length) {
-      // Shift everything over to make room
-      System.arraycopy(array, index, array, index + 1, size - index);
-    } else {
-      // Resize to 1.5x the size
-      int length = ((size * 3) / 2) + 1;
-      boolean[] newArray = new boolean[length];
-
-      // Copy the first part directly
-      System.arraycopy(array, 0, newArray, 0, index);
-
-      // Copy the rest shifted over by one to make room
-      System.arraycopy(array, index, newArray, index + 1, size - index);
-      array = newArray;
-    }
-
-    array[index] = element;
-    size++;
-    modCount++;
-  }
-
-  @Override
-  public boolean addAll(Collection<? extends Boolean> collection) {
-    ensureIsMutable();
-
-    if (collection == null) {
-      throw new NullPointerException();
-    }
-
-    // We specialize when adding another BooleanArrayList to avoid boxing elements.
-    if (!(collection instanceof BooleanArrayList)) {
-      return super.addAll(collection);
-    }
-
-    BooleanArrayList list = (BooleanArrayList) collection;
-    if (list.size == 0) {
-      return false;
-    }
-
-    int overflow = Integer.MAX_VALUE - size;
-    if (overflow < list.size) {
-      // We can't actually represent a list this large.
-      throw new OutOfMemoryError();
-    }
-
-    int newSize = size + list.size;
-    if (newSize > array.length) {
-      array = Arrays.copyOf(array, newSize);
-    }
-
-    System.arraycopy(list.array, 0, array, size, list.size);
-    size = newSize;
-    modCount++;
-    return true;
-  }
-
-  @Override
-  public boolean remove(Object o) {
-    ensureIsMutable();
-    for (int i = 0; i < size; i++) {
-      if (o.equals(array[i])) {
-        System.arraycopy(array, i + 1, array, i, size - i);
-        size--;
-        modCount++;
-        return true;
-      }
-    }
-    return false;
-  }
-
-  @Override
-  public Boolean remove(int index) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    boolean value = array[index];
-    System.arraycopy(array, index + 1, array, index, size - index);
-    size--;
-    modCount++;
-    return value;
-  }
-
-  /**
-   * Ensures that the provided {@code index} is within the range of {@code [0, size]}. Throws an
-   * {@link IndexOutOfBoundsException} if it is not.
-   *
-   * @param index the index to verify is in range
-   */
-  private void ensureIndexInRange(int index) {
-    if (index < 0 || index >= size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-  }
-
-  private String makeOutOfBoundsExceptionMessage(int index) {
-    return "Index:" + index + ", Size:" + size;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java
deleted file mode 100644
index 553243a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java
+++ /dev/null
@@ -1,185 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import static java.lang.Math.max;
-import static java.lang.Math.min;
-
-import java.io.IOException;
-import java.io.OutputStream;
-import java.lang.ref.SoftReference;
-import java.lang.reflect.Field;
-import java.nio.ByteBuffer;
-import java.nio.channels.WritableByteChannel;
-
-/**
- * Utility class to provide efficient writing of {@link ByteBuffer}s to {@link OutputStream}s.
- */
-final class ByteBufferWriter {
-  private ByteBufferWriter() {}
-
-  /**
-   * Minimum size for a cached buffer. This prevents us from allocating buffers that are too
-   * small to be easily reused.
-   */
-  // TODO(nathanmittler): tune this property or allow configuration?
-  private static final int MIN_CACHED_BUFFER_SIZE = 1024;
-
-  /**
-   * Maximum size for a cached buffer. If a larger buffer is required, it will be allocated
-   * but not cached.
-   */
-  // TODO(nathanmittler): tune this property or allow configuration?
-  private static final int MAX_CACHED_BUFFER_SIZE = 16 * 1024;
-
-  /**
-   * The fraction of the requested buffer size under which the buffer will be reallocated.
-   */
-  // TODO(nathanmittler): tune this property or allow configuration?
-  private static final float BUFFER_REALLOCATION_THRESHOLD = 0.5f;
-
-  /**
-   * Keeping a soft reference to a thread-local buffer. This buffer is used for writing a
-   * {@link ByteBuffer} to an {@link OutputStream} when no zero-copy alternative was available.
-   * Using a "soft" reference since VMs may keep this reference around longer than "weak"
-   * (e.g. HotSpot will maintain soft references until memory pressure warrants collection).
-   */
-  private static final ThreadLocal<SoftReference<byte[]>> BUFFER =
-      new ThreadLocal<SoftReference<byte[]>>();
-
-  /**
-   * This is a hack for GAE, where {@code FileOutputStream} is unavailable.
-   */
-  private static final Class<?> FILE_OUTPUT_STREAM_CLASS = safeGetClass("java.io.FileOutputStream");
-  private static final long CHANNEL_FIELD_OFFSET = getChannelFieldOffset(FILE_OUTPUT_STREAM_CLASS);
-
-  /**
-   * For testing purposes only. Clears the cached buffer to force a new allocation on the next
-   * invocation.
-   */
-  static void clearCachedBuffer() {
-    BUFFER.set(null);
-  }
-
-  /**
-   * Writes the remaining content of the buffer to the given stream. The buffer {@code position}
-   * will remain unchanged by this method.
-   */
-  static void write(ByteBuffer buffer, OutputStream output) throws IOException {
-    final int initialPos = buffer.position();
-    try {
-      if (buffer.hasArray()) {
-        // Optimized write for array-backed buffers.
-        // Note that we're taking the risk that a malicious OutputStream could modify the array.
-        output.write(buffer.array(), buffer.arrayOffset() + buffer.position(), buffer.remaining());
-      } else if (!writeToChannel(buffer, output)){
-        // Read all of the data from the buffer to an array.
-        // TODO(nathanmittler): Consider performance improvements for other "known" stream types.
-        final byte[] array = getOrCreateBuffer(buffer.remaining());
-        while (buffer.hasRemaining()) {
-          int length = min(buffer.remaining(), array.length);
-          buffer.get(array, 0, length);
-          output.write(array, 0, length);
-        }
-      }
-    } finally {
-      // Restore the initial position.
-      buffer.position(initialPos);
-    }
-  }
-
-  static byte[] getOrCreateBuffer(int requestedSize) {
-    requestedSize = max(requestedSize, MIN_CACHED_BUFFER_SIZE);
-
-    byte[] buffer = getBuffer();
-    // Only allocate if we need to.
-    if (buffer == null || needToReallocate(requestedSize, buffer.length)) {
-      buffer = new byte[requestedSize];
-
-      // Only cache the buffer if it's not too big.
-      if (requestedSize <= MAX_CACHED_BUFFER_SIZE) {
-        setBuffer(buffer);
-      }
-    }
-    return buffer;
-  }
-
-  private static boolean needToReallocate(int requestedSize, int bufferLength) {
-    // First check against just the requested length to avoid the multiply.
-    return bufferLength < requestedSize
-        && bufferLength < requestedSize * BUFFER_REALLOCATION_THRESHOLD;
-  }
-
-  private static byte[] getBuffer() {
-    SoftReference<byte[]> sr = BUFFER.get();
-    return sr == null ? null : sr.get();
-  }
-
-  private static void setBuffer(byte[] value) {
-    BUFFER.set(new SoftReference<byte[]>(value));
-  }
-
-  private static boolean writeToChannel(ByteBuffer buffer, OutputStream output) throws IOException {
-    if (CHANNEL_FIELD_OFFSET >= 0 && FILE_OUTPUT_STREAM_CLASS.isInstance(output)) {
-      // Use a channel to write out the ByteBuffer. This will automatically empty the buffer.
-      WritableByteChannel channel = null;
-      try {
-        channel = (WritableByteChannel) UnsafeUtil.getObject(output, CHANNEL_FIELD_OFFSET);
-      } catch (ClassCastException e) {
-        // Absorb.
-      }
-      if (channel != null) {
-        channel.write(buffer);
-        return true;
-      }
-    }
-    return false;
-  }
-
-  private static Class<?> safeGetClass(String className) {
-    try {
-      return Class.forName(className);
-    } catch (ClassNotFoundException e) {
-      return null;
-    }
-  }
-  private static long getChannelFieldOffset(Class<?> clazz) {
-    try {
-      if (clazz != null && UnsafeUtil.hasUnsafeArrayOperations()) {
-        Field field = clazz.getDeclaredField("channel");
-        return UnsafeUtil.objectFieldOffset(field);
-      }
-    } catch (Throwable e) {
-      // Absorb
-    }
-    return -1;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java
deleted file mode 100644
index a745d37..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java
+++ /dev/null
@@ -1,81 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.nio.ByteBuffer;
-
-/**
- * An input for raw bytes. This is similar to an InputStream but it is offset addressable. All the
- * read APIs are relative.
- */
-@ExperimentalApi
-public abstract class ByteInput {
-
-  /**
-   * Reads a single byte from the given offset.
-   * @param offset The offset from where byte to be read
-   * @return The byte of data at given offset
-   */
-  public abstract byte read(int offset);
-
-  /**
-   * Reads bytes of data from the given offset into an array of bytes.
-   * @param offset The src offset within this ByteInput from where data to be read.
-   * @param out Destination byte array to read data into.
-   * @return The number of bytes read from ByteInput
-   */
-  public int read(int offset, byte b[]) throws IOException {
-    return read(offset, b, 0, b.length);
-  }
-
-  /**
-   * Reads up to <code>len</code> bytes of data from the given offset into an array of bytes.
-   * @param offset The src offset within this ByteInput from where data to be read.
-   * @param out Destination byte array to read data into.
-   * @param outOffset Offset within the the out byte[] where data to be read into.
-   * @param len The number of bytes to read.
-   * @return The number of bytes read from ByteInput
-   */
-  public abstract int read(int offset, byte[] out, int outOffset, int len);
-
-  /**
-   * Reads bytes of data from the given offset into given {@link ByteBuffer}.
-   * @param offset he src offset within this ByteInput from where data to be read.
-   * @param out Destination {@link ByteBuffer} to read data into.
-   * @return The number of bytes read from ByteInput
-   */
-  public abstract int read(int offset, ByteBuffer out);
-
-  /**
-   * @return Total number of bytes in this ByteInput.
-   */
-  public abstract int size();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java
deleted file mode 100644
index 30de4ec..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java
+++ /dev/null
@@ -1,251 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.InvalidObjectException;
-import java.io.ObjectInputStream;
-import java.io.OutputStream;
-import java.nio.ByteBuffer;
-import java.nio.charset.Charset;
-import java.util.Arrays;
-import java.util.Collections;
-import java.util.List;
-
-/**
- * A {@link ByteString} that wraps around a {@link ByteInput}.
- */
-final class ByteInputByteString extends ByteString.LeafByteString {
-  private final ByteInput buffer;
-  private final int offset, length;
-
-  ByteInputByteString(ByteInput buffer, int offset, int length) {
-    if (buffer == null) {
-      throw new NullPointerException("buffer");
-    }
-    this.buffer = buffer;
-    this.offset = offset;
-    this.length = length;
-  }
-
-  // =================================================================
-  // Serializable
-
-  /**
-   * Magic method that lets us override serialization behavior.
-   */
-  private Object writeReplace() {
-    return ByteString.wrap(toByteArray());
-  }
-
-  /**
-   * Magic method that lets us override deserialization behavior.
-   */
-  private void readObject(@SuppressWarnings("unused") ObjectInputStream in) throws IOException {
-    throw new InvalidObjectException("ByteInputByteString instances are not to be serialized directly");// TODO check here
-  }
-
-  // =================================================================
-
-  @Override
-  public byte byteAt(int index) {
-    return buffer.read(getAbsoluteOffset(index));
-  }
-
-  private int getAbsoluteOffset(int relativeOffset) {
-    return this.offset + relativeOffset;
-  }
-
-  @Override
-  public int size() {
-    return length;
-  }
-
-  @Override
-  public ByteString substring(int beginIndex, int endIndex) {
-    if (beginIndex < 0 || beginIndex >= size() || endIndex < beginIndex || endIndex >= size()) {
-      throw new IllegalArgumentException(
-          String.format("Invalid indices [%d, %d]", beginIndex, endIndex));
-    }
-    return new ByteInputByteString(this.buffer, getAbsoluteOffset(beginIndex), endIndex - beginIndex);
-  }
-
-  @Override
-  protected void copyToInternal(
-      byte[] target, int sourceOffset, int targetOffset, int numberToCopy) {
-    this.buffer.read(getAbsoluteOffset(sourceOffset), target, targetOffset, numberToCopy);
-  }
-
-  @Override
-  public void copyTo(ByteBuffer target) {
-    this.buffer.read(this.offset, target);
-  }
-
-  @Override
-  public void writeTo(OutputStream out) throws IOException {
-    out.write(toByteArray());// TODO
-  }
-
-  @Override
-  boolean equalsRange(ByteString other, int offset, int length) {
-    return substring(0, length).equals(other.substring(offset, offset + length));
-  }
-
-  @Override
-  void writeToInternal(OutputStream out, int sourceOffset, int numberToWrite) throws IOException {
-    byte[] buf = ByteBufferWriter.getOrCreateBuffer(numberToWrite);
-    this.buffer.read(getAbsoluteOffset(sourceOffset), buf, 0, numberToWrite);
-    out.write(buf, 0, numberToWrite);
-  }
-
-  @Override
-  void writeTo(ByteOutput output) throws IOException {
-    output.writeLazy(toByteArray(), 0, length);
-  }
-
-  @Override
-  public ByteBuffer asReadOnlyByteBuffer() {
-    return ByteBuffer.wrap(toByteArray()).asReadOnlyBuffer();
-  }
-
-  @Override
-  public List<ByteBuffer> asReadOnlyByteBufferList() {
-    return Collections.singletonList(asReadOnlyByteBuffer());
-  }
-
-  @Override
-  protected String toStringInternal(Charset charset) {
-    byte[] bytes = toByteArray();
-    return new String(bytes, 0, bytes.length, charset);
-  }
-
-  @Override
-  public boolean isValidUtf8() {
-    return Utf8.isValidUtf8(buffer, offset, offset + length);
-  }
-
-  @Override
-  protected int partialIsValidUtf8(int state, int offset, int length) {
-    int off = getAbsoluteOffset(offset);
-    return Utf8.partialIsValidUtf8(state, buffer, off, off + length);
-  }
-
-  @Override
-  public boolean equals(Object other) {
-    if (other == this) {
-      return true;
-    }
-    if (!(other instanceof ByteString)) {
-      return false;
-    }
-    ByteString otherString = ((ByteString) other);
-    if (size() != otherString.size()) {
-      return false;
-    }
-    if (size() == 0) {
-      return true;
-    }
-    if (other instanceof RopeByteString) {
-      return other.equals(this);
-    }
-    return Arrays.equals(this.toByteArray(), otherString.toByteArray());
-  }
-
-  @Override
-  protected int partialHash(int h, int offset, int length) {
-    offset = getAbsoluteOffset(offset);
-    int end = offset + length;
-    for (int i = offset; i < end; i++) {
-      h = h * 31 + buffer.read(i);
-    }
-    return h;
-  }
-
-  @Override
-  public InputStream newInput() {
-    return new InputStream() {
-      private final ByteInput buf = buffer;
-      private int pos = offset;
-      private int limit = pos + length;
-      private int mark = pos;
-
-      @Override
-      public void mark(int readlimit) {
-        this.mark = readlimit;
-      }
-
-      @Override
-      public boolean markSupported() {
-        return true;
-      }
-
-      @Override
-      public void reset() throws IOException {
-        this.pos = this.mark;
-      }
-
-      @Override
-      public int available() throws IOException {
-        return this.limit - this.pos;
-      }
-
-      @Override
-      public int read() throws IOException {
-        if (available() <= 0) {
-          return -1;
-        }
-        return this.buf.read(pos++) & 0xFF;
-      }
-
-      @Override
-      public int read(byte[] bytes, int off, int len) throws IOException {
-        int remain = available();
-        if (remain <= 0) {
-          return -1;
-        }
-        len = Math.min(len, remain);
-        buf.read(pos, bytes, off, len);
-        pos += len;
-        return len;
-      }
-    };
-  }
-
-  @Override
-  public CodedInputStream newCodedInput() {
-    // We trust CodedInputStream not to modify the bytes, or to give anyone
-    // else access to them.
-    CodedInputStream cis = CodedInputStream.newInstance(buffer, offset, length, true);
-    cis.enableAliasing(true);
-    return cis;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java
deleted file mode 100644
index 8a83fb2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java
+++ /dev/null
@@ -1,116 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.nio.ByteBuffer;
-
-/**
- * An output target for raw bytes. This interface provides semantics that support two types of
- * writing:
- *
- * <p><b>Traditional write operations:</b>
- * (as defined by {@link java.io.OutputStream}) where the target method is responsible for either
- * copying the data or completing the write before returning from the method call.
- *
- * <p><b>Lazy write operations:</b> where the caller guarantees that it will never modify the
- * provided buffer and it can therefore be considered immutable. The target method is free to
- * maintain a reference to the buffer beyond the scope of the method call (e.g. until the write
- * operation completes).
- */
-@ExperimentalApi
-public abstract class ByteOutput {
-  /**
-   * Writes a single byte.
-   *
-   * @param value the byte to be written
-   * @throws IOException thrown if an error occurred while writing
-   */
-  public abstract void write(byte value) throws IOException;
-
-  /**
-   * Writes a sequence of bytes. The {@link ByteOutput} must copy {@code value} if it will
-   * not be processed prior to the return of this method call, since {@code value} may be
-   * reused/altered by the caller.
-   *
-   * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a
-   * programming error and will lead to data corruption which will be difficult to debug.
-   *
-   * @param value the bytes to be written
-   * @param offset the offset of the start of the writable range
-   * @param length the number of bytes to write starting from {@code offset}
-   * @throws IOException thrown if an error occurred while writing
-   */
-  public abstract void write(byte[] value, int offset, int length) throws IOException;
-
-  /**
-   * Writes a sequence of bytes. The {@link ByteOutput} is free to retain a reference to the value
-   * beyond the scope of this method call (e.g. write later) since it is considered immutable and is
-   * guaranteed not to change by the caller.
-   *
-   * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a
-   * programming error and will lead to data corruption which will be difficult to debug.
-   *
-   * @param value the bytes to be written
-   * @param offset the offset of the start of the writable range
-   * @param length the number of bytes to write starting from {@code offset}
-   * @throws IOException thrown if an error occurred while writing
-   */
-  public abstract void writeLazy(byte[] value, int offset, int length) throws IOException;
-
-  /**
-   * Writes a sequence of bytes. The {@link ByteOutput} must copy {@code value} if it will
-   * not be processed prior to the return of this method call, since {@code value} may be
-   * reused/altered by the caller.
-   *
-   * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a
-   * programming error and will lead to data corruption which will be difficult to debug.
-   *
-   * @param value the bytes to be written. Upon returning from this call, the {@code position} of
-   * this buffer will be set to the {@code limit}
-   * @throws IOException thrown if an error occurred while writing
-   */
-  public abstract void write(ByteBuffer value) throws IOException;
-
-  /**
-   * Writes a sequence of bytes. The {@link ByteOutput} is free to retain a reference to the value
-   * beyond the scope of this method call (e.g. write later) since it is considered immutable and is
-   * guaranteed not to change by the caller.
-   *
-   * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a
-   * programming error and will lead to data corruption which will be difficult to debug.
-   *
-   * @param value the bytes to be written. Upon returning from this call, the {@code position} of
-   * this buffer will be set to the {@code limit}
-   * @throws IOException thrown if an error occurred while writing
-   */
-  public abstract void writeLazy(ByteBuffer value) throws IOException;
-}


[09/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Utf8.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Utf8.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Utf8.java
deleted file mode 100644
index b84efd6..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Utf8.java
+++ /dev/null
@@ -1,1764 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeUtil.addressOffset;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeUtil.getArrayBaseOffset;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeUtil.hasUnsafeArrayOperations;
-import static org.apache.hadoop.hbase.shaded.com.google.protobuf.UnsafeUtil.hasUnsafeByteBufferOperations;
-import static java.lang.Character.MAX_SURROGATE;
-import static java.lang.Character.MIN_SURROGATE;
-import static java.lang.Character.isSurrogatePair;
-import static java.lang.Character.toCodePoint;
-
-import java.nio.ByteBuffer;
-
-/**
- * A set of low-level, high-performance static utility methods related
- * to the UTF-8 character encoding.  This class has no dependencies
- * outside of the core JDK libraries.
- *
- * <p>There are several variants of UTF-8.  The one implemented by
- * this class is the restricted definition of UTF-8 introduced in
- * Unicode 3.1, which mandates the rejection of "overlong" byte
- * sequences as well as rejection of 3-byte surrogate codepoint byte
- * sequences.  Note that the UTF-8 decoder included in Oracle's JDK
- * has been modified to also reject "overlong" byte sequences, but (as
- * of 2011) still accepts 3-byte surrogate codepoint byte sequences.
- *
- * <p>The byte sequences considered valid by this class are exactly
- * those that can be roundtrip converted to Strings and back to bytes
- * using the UTF-8 charset, without loss: <pre> {@code
- * Arrays.equals(bytes, new String(bytes, Internal.UTF_8).getBytes(Internal.UTF_8))
- * }</pre>
- *
- * <p>See the Unicode Standard,</br>
- * Table 3-6. <em>UTF-8 Bit Distribution</em>,</br>
- * Table 3-7. <em>Well Formed UTF-8 Byte Sequences</em>.
- *
- * <p>This class supports decoding of partial byte sequences, so that the
- * bytes in a complete UTF-8 byte sequences can be stored in multiple
- * segments.  Methods typically return {@link #MALFORMED} if the partial
- * byte sequence is definitely not well-formed, {@link #COMPLETE} if it is
- * well-formed in the absence of additional input, or if the byte sequence
- * apparently terminated in the middle of a character, an opaque integer
- * "state" value containing enough information to decode the character when
- * passed to a subsequent invocation of a partial decoding method.
- *
- * @author martinrb@google.com (Martin Buchholz)
- */
-// TODO(nathanmittler): Copy changes in this class back to Guava
-final class Utf8 {
-
-  /**
-   * UTF-8 is a runtime hot spot so we attempt to provide heavily optimized implementations
-   * depending on what is available on the platform. The processor is the platform-optimized
-   * delegate for which all methods are delegated directly to.
-   */
-  private static final Processor processor =
-      UnsafeProcessor.isAvailable() ? new UnsafeProcessor() : new SafeProcessor();
-
-  /**
-   * A mask used when performing unsafe reads to determine if a long value contains any non-ASCII
-   * characters (i.e. any byte >= 0x80).
-   */
-  private static final long ASCII_MASK_LONG = 0x8080808080808080L;
-
-  /**
-   * Maximum number of bytes per Java UTF-16 char in UTF-8.
-   * @see java.nio.charset.CharsetEncoder#maxBytesPerChar()
-   */
-  static final int MAX_BYTES_PER_CHAR = 3;
-
-  /**
-   * State value indicating that the byte sequence is well-formed and
-   * complete (no further bytes are needed to complete a character).
-   */
-  public static final int COMPLETE = 0;
-
-  /**
-   * State value indicating that the byte sequence is definitely not
-   * well-formed.
-   */
-  public static final int MALFORMED = -1;
-
-  /**
-   * Used by {@code Unsafe} UTF-8 string validation logic to determine the minimum string length
-   * above which to employ an optimized algorithm for counting ASCII characters. The reason for this
-   * threshold is that for small strings, the optimization may not be beneficial or may even
-   * negatively impact performance since it requires additional logic to avoid unaligned reads
-   * (when calling {@code Unsafe.getLong}). This threshold guarantees that even if the initial
-   * offset is unaligned, we're guaranteed to make at least one call to {@code Unsafe.getLong()}
-   * which provides a performance improvement that entirely subsumes the cost of the additional
-   * logic.
-   */
-  private static final int UNSAFE_COUNT_ASCII_THRESHOLD = 16;
-
-  // Other state values include the partial bytes of the incomplete
-  // character to be decoded in the simplest way: we pack the bytes
-  // into the state int in little-endian order.  For example:
-  //
-  // int state = byte1 ^ (byte2 << 8) ^ (byte3 << 16);
-  //
-  // Such a state is unpacked thus (note the ~ operation for byte2 to
-  // undo byte1's sign-extension bits):
-  //
-  // int byte1 = (byte) state;
-  // int byte2 = (byte) ~(state >> 8);
-  // int byte3 = (byte) (state >> 16);
-  //
-  // We cannot store a zero byte in the state because it would be
-  // indistinguishable from the absence of a byte.  But we don't need
-  // to, because partial bytes must always be negative.  When building
-  // a state, we ensure that byte1 is negative and subsequent bytes
-  // are valid trailing bytes.
-
-  /**
-   * Returns {@code true} if the given byte array is a well-formed
-   * UTF-8 byte sequence.
-   *
-   * <p>This is a convenience method, equivalent to a call to {@code
-   * isValidUtf8(bytes, 0, bytes.length)}.
-   */
-  public static boolean isValidUtf8(byte[] bytes) {
-    return processor.isValidUtf8(bytes, 0, bytes.length);
-  }
-
-  /**
-   * Returns {@code true} if the given byte array slice is a
-   * well-formed UTF-8 byte sequence.  The range of bytes to be
-   * checked extends from index {@code index}, inclusive, to {@code
-   * limit}, exclusive.
-   *
-   * <p>This is a convenience method, equivalent to {@code
-   * partialIsValidUtf8(bytes, index, limit) == Utf8.COMPLETE}.
-   */
-  public static boolean isValidUtf8(byte[] bytes, int index, int limit) {
-    return processor.isValidUtf8(bytes, index, limit);
-  }
-
-  /**
-   * Tells whether the given byte array slice is a well-formed,
-   * malformed, or incomplete UTF-8 byte sequence.  The range of bytes
-   * to be checked extends from index {@code index}, inclusive, to
-   * {@code limit}, exclusive.
-   *
-   * @param state either {@link Utf8#COMPLETE} (if this is the initial decoding
-   * operation) or the value returned from a call to a partial decoding method
-   * for the previous bytes
-   *
-   * @return {@link #MALFORMED} if the partial byte sequence is
-   * definitely not well-formed, {@link #COMPLETE} if it is well-formed
-   * (no additional input needed), or if the byte sequence is
-   * "incomplete", i.e. apparently terminated in the middle of a character,
-   * an opaque integer "state" value containing enough information to
-   * decode the character when passed to a subsequent invocation of a
-   * partial decoding method.
-   */
-  public static int partialIsValidUtf8(int state, byte[] bytes, int index, int limit) {
-    return processor.partialIsValidUtf8(state, bytes, index, limit);
-  }
-
-  private static int incompleteStateFor(int byte1) {
-    return (byte1 > (byte) 0xF4) ?
-        MALFORMED : byte1;
-  }
-
-  private static int incompleteStateFor(int byte1, int byte2) {
-    return (byte1 > (byte) 0xF4 ||
-            byte2 > (byte) 0xBF) ?
-        MALFORMED : byte1 ^ (byte2 << 8);
-  }
-
-  private static int incompleteStateFor(int byte1, int byte2, int byte3) {
-    return (byte1 > (byte) 0xF4 ||
-            byte2 > (byte) 0xBF ||
-            byte3 > (byte) 0xBF) ?
-        MALFORMED : byte1 ^ (byte2 << 8) ^ (byte3 << 16);
-  }
-
-  private static int incompleteStateFor(byte[] bytes, int index, int limit) {
-    int byte1 = bytes[index - 1];
-    switch (limit - index) {
-      case 0: return incompleteStateFor(byte1);
-      case 1: return incompleteStateFor(byte1, bytes[index]);
-      case 2: return incompleteStateFor(byte1, bytes[index], bytes[index + 1]);
-      default: throw new AssertionError();
-    }
-  }
-
-  private static int incompleteStateFor(
-      final ByteBuffer buffer, final int byte1, final int index, final int remaining) {
-    switch (remaining) {
-      case 0:
-        return incompleteStateFor(byte1);
-      case 1:
-        return incompleteStateFor(byte1, buffer.get(index));
-      case 2:
-        return incompleteStateFor(byte1, buffer.get(index), buffer.get(index + 1));
-      default:
-        throw new AssertionError();
-    }
-  }
-
-  private static int incompleteStateFor(ByteInput bytes, int index, int limit) {
-    int byte1 = bytes.read(index - 1);
-    switch (limit - index) {
-      case 0: return incompleteStateFor(byte1);
-      case 1: return incompleteStateFor(byte1, bytes.read(index));
-      case 2: return incompleteStateFor(byte1, bytes.read(index), bytes.read(index + 1));
-      default: throw new AssertionError();
-    }
-  }
-
-  // These UTF-8 handling methods are copied from Guava's Utf8 class with a modification to throw
-  // a protocol buffer local exception. This exception is then caught in CodedOutputStream so it can
-  // fallback to more lenient behavior.
-  
-  static class UnpairedSurrogateException extends IllegalArgumentException {
-    UnpairedSurrogateException(int index, int length) {
-      super("Unpaired surrogate at index " + index + " of " + length);
-    }
-  }
-
-  /**
-   * Returns the number of bytes in the UTF-8-encoded form of {@code sequence}. For a string,
-   * this method is equivalent to {@code string.getBytes(UTF_8).length}, but is more efficient in
-   * both time and space.
-   *
-   * @throws IllegalArgumentException if {@code sequence} contains ill-formed UTF-16 (unpaired
-   *     surrogates)
-   */
-  static int encodedLength(CharSequence sequence) {
-    // Warning to maintainers: this implementation is highly optimized.
-    int utf16Length = sequence.length();
-    int utf8Length = utf16Length;
-    int i = 0;
-
-    // This loop optimizes for pure ASCII.
-    while (i < utf16Length && sequence.charAt(i) < 0x80) {
-      i++;
-    }
-
-    // This loop optimizes for chars less than 0x800.
-    for (; i < utf16Length; i++) {
-      char c = sequence.charAt(i);
-      if (c < 0x800) {
-        utf8Length += ((0x7f - c) >>> 31);  // branch free!
-      } else {
-        utf8Length += encodedLengthGeneral(sequence, i);
-        break;
-      }
-    }
-
-    if (utf8Length < utf16Length) {
-      // Necessary and sufficient condition for overflow because of maximum 3x expansion
-      throw new IllegalArgumentException("UTF-8 length does not fit in int: "
-              + (utf8Length + (1L << 32)));
-    }
-    return utf8Length;
-  }
-
-  private static int encodedLengthGeneral(CharSequence sequence, int start) {
-    int utf16Length = sequence.length();
-    int utf8Length = 0;
-    for (int i = start; i < utf16Length; i++) {
-      char c = sequence.charAt(i);
-      if (c < 0x800) {
-        utf8Length += (0x7f - c) >>> 31; // branch free!
-      } else {
-        utf8Length += 2;
-        // jdk7+: if (Character.isSurrogate(c)) {
-        if (Character.MIN_SURROGATE <= c && c <= Character.MAX_SURROGATE) {
-          // Check that we have a well-formed surrogate pair.
-          int cp = Character.codePointAt(sequence, i);
-          if (cp < Character.MIN_SUPPLEMENTARY_CODE_POINT) {
-            throw new UnpairedSurrogateException(i, utf16Length);
-          }
-          i++;
-        }
-      }
-    }
-    return utf8Length;
-  }
-
-  static int encode(CharSequence in, byte[] out, int offset, int length) {
-    return processor.encodeUtf8(in, out, offset, length);
-  }
-  // End Guava UTF-8 methods.
-
-  /**
-   * Determines if the given {@link ByteBuffer} is a valid UTF-8 string.
-   *
-   * <p>Selects an optimal algorithm based on the type of {@link ByteBuffer} (i.e. heap or direct)
-   * and the capabilities of the platform.
-   *
-   * @param buffer the buffer to check.
-   * @see Utf8#isValidUtf8(byte[], int, int)
-   */
-  static boolean isValidUtf8(ByteBuffer buffer) {
-    return processor.isValidUtf8(buffer, buffer.position(), buffer.remaining());
-  }
-
-  /**
-   * Determines if the given {@link ByteBuffer} is a partially valid UTF-8 string.
-   *
-   * <p>Selects an optimal algorithm based on the type of {@link ByteBuffer} (i.e. heap or direct)
-   * and the capabilities of the platform.
-   *
-   * @param buffer the buffer to check.
-   * @see Utf8#partialIsValidUtf8(int, byte[], int, int)
-   */
-  static int partialIsValidUtf8(int state, ByteBuffer buffer, int index, int limit) {
-    return processor.partialIsValidUtf8(state, buffer, index, limit);
-  }
-
-  /**
-   * Determines if the given {@link ByteInput} is a valid UTF-8 string.
-   *
-   * @param buffer the buffer to check.
-   */
-  static boolean isValidUtf8(ByteInput buffer, int index, int limit) {
-    return processor.isValidUtf8(buffer, index, limit);
-  }
-
-  /**
-   * Determines if the given {@link ByteInput} is a partially valid UTF-8 string.
-   *
-   * @param buffer the buffer to check.
-   */
-  static int partialIsValidUtf8(int state, ByteInput buffer, int index, int limit) {
-    return processor.partialIsValidUtf8(state, buffer, index, limit);
-  }
-
-  /**
-   * Encodes the given characters to the target {@link ByteBuffer} using UTF-8 encoding.
-   *
-   * <p>Selects an optimal algorithm based on the type of {@link ByteBuffer} (i.e. heap or direct)
-   * and the capabilities of the platform.
-   *
-   * @param in the source string to be encoded
-   * @param out the target buffer to receive the encoded string.
-   * @see Utf8#encode(CharSequence, byte[], int, int)
-   */
-  static void encodeUtf8(CharSequence in, ByteBuffer out) {
-    processor.encodeUtf8(in, out);
-  }
-
-  /**
-   * Counts (approximately) the number of consecutive ASCII characters in the given buffer.
-   * The byte order of the {@link ByteBuffer} does not matter, so performance can be improved if
-   * native byte order is used (i.e. no byte-swapping in {@link ByteBuffer#getLong(int)}).
-   *
-   * @param buffer the buffer to be scanned for ASCII chars
-   * @param index the starting index of the scan
-   * @param limit the limit within buffer for the scan
-   * @return the number of ASCII characters found. The stopping position will be at or
-   * before the first non-ASCII byte.
-   */
-  private static int estimateConsecutiveAscii(ByteBuffer buffer, int index, int limit) {
-    int i = index;
-    final int lim = limit - 7;
-    // This simple loop stops when we encounter a byte >= 0x80 (i.e. non-ASCII).
-    // To speed things up further, we're reading longs instead of bytes so we use a mask to
-    // determine if any byte in the current long is non-ASCII.
-    for (; i < lim && (buffer.getLong(i) & ASCII_MASK_LONG) == 0; i += 8) {}
-    return i - index;
-  }
-
-  /**
-   * A processor of UTF-8 strings, providing methods for checking validity and encoding.
-   */
-  // TODO(nathanmittler): Add support for Memory/MemoryBlock on Android.
-  abstract static class Processor {
-    /**
-     * Returns {@code true} if the given byte array slice is a
-     * well-formed UTF-8 byte sequence.  The range of bytes to be
-     * checked extends from index {@code index}, inclusive, to {@code
-     * limit}, exclusive.
-     *
-     * <p>This is a convenience method, equivalent to {@code
-     * partialIsValidUtf8(bytes, index, limit) == Utf8.COMPLETE}.
-     */
-    final boolean isValidUtf8(byte[] bytes, int index, int limit) {
-      return partialIsValidUtf8(COMPLETE, bytes, index, limit) == COMPLETE;
-    }
-
-    /**
-     * Tells whether the given byte array slice is a well-formed,
-     * malformed, or incomplete UTF-8 byte sequence.  The range of bytes
-     * to be checked extends from index {@code index}, inclusive, to
-     * {@code limit}, exclusive.
-     *
-     * @param state either {@link Utf8#COMPLETE} (if this is the initial decoding
-     * operation) or the value returned from a call to a partial decoding method
-     * for the previous bytes
-     *
-     * @return {@link #MALFORMED} if the partial byte sequence is
-     * definitely not well-formed, {@link #COMPLETE} if it is well-formed
-     * (no additional input needed), or if the byte sequence is
-     * "incomplete", i.e. apparently terminated in the middle of a character,
-     * an opaque integer "state" value containing enough information to
-     * decode the character when passed to a subsequent invocation of a
-     * partial decoding method.
-     */
-    abstract int partialIsValidUtf8(int state, byte[] bytes, int index, int limit);
-
-    /**
-     * Returns {@code true} if the given portion of the {@link ByteBuffer} is a
-     * well-formed UTF-8 byte sequence.  The range of bytes to be
-     * checked extends from index {@code index}, inclusive, to {@code
-     * limit}, exclusive.
-     *
-     * <p>This is a convenience method, equivalent to {@code
-     * partialIsValidUtf8(bytes, index, limit) == Utf8.COMPLETE}.
-     */
-    final boolean isValidUtf8(ByteBuffer buffer, int index, int limit) {
-      return partialIsValidUtf8(COMPLETE, buffer, index, limit) == COMPLETE;
-    }
-
-    /**
-     * Indicates whether or not the given buffer contains a valid UTF-8 string.
-     *
-     * @param buffer the buffer to check.
-     * @return {@code true} if the given buffer contains a valid UTF-8 string.
-     */
-    final int partialIsValidUtf8(
-        final int state, final ByteBuffer buffer, int index, final int limit) {
-      if (buffer.hasArray()) {
-        final int offset = buffer.arrayOffset();
-        return partialIsValidUtf8(state, buffer.array(), offset + index, offset + limit);
-      } else if (buffer.isDirect()){
-        return partialIsValidUtf8Direct(state, buffer, index, limit);
-      }
-      return partialIsValidUtf8Default(state, buffer, index, limit);
-    }
-
-    /**
-     * Performs validation for direct {@link ByteBuffer} instances.
-     */
-    abstract int partialIsValidUtf8Direct(
-        final int state, final ByteBuffer buffer, int index, final int limit);
-
-    /**
-     * Performs validation for {@link ByteBuffer} instances using the {@link ByteBuffer} API rather
-     * than potentially faster approaches. This first completes validation for the current
-     * character (provided by {@code state}) and then finishes validation for the sequence.
-     */
-    final int partialIsValidUtf8Default(
-        final int state, final ByteBuffer buffer, int index, final int limit) {
-      if (state != COMPLETE) {
-        // The previous decoding operation was incomplete (or malformed).
-        // We look for a well-formed sequence consisting of bytes from
-        // the previous decoding operation (stored in state) together
-        // with bytes from the array slice.
-        //
-        // We expect such "straddler characters" to be rare.
-
-        if (index >= limit) { // No bytes? No progress.
-          return state;
-        }
-
-        byte byte1 = (byte) state;
-        // byte1 is never ASCII.
-        if (byte1 < (byte) 0xE0) {
-          // two-byte form
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2
-              // byte2 trailing-byte test
-              || buffer.get(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // three-byte form
-
-          // Get byte2 from saved state or array
-          byte byte2 = (byte) ~(state >> 8);
-          if (byte2 == 0) {
-            byte2 = buffer.get(index++);
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          }
-          if (byte2 > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // illegal surrogate codepoint?
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || buffer.get(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // four-byte form
-
-          // Get byte2 and byte3 from saved state or array
-          byte byte2 = (byte) ~(state >> 8);
-          byte byte3 = 0;
-          if (byte2 == 0) {
-            byte2 = buffer.get(index++);
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          } else {
-            byte3 = (byte) (state >> 16);
-          }
-          if (byte3 == 0) {
-            byte3 = buffer.get(index++);
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2, byte3);
-            }
-          }
-
-          // If we were called with state == MALFORMED, then byte1 is 0xFF,
-          // which never occurs in well-formed UTF-8, and so we will return
-          // MALFORMED again below.
-
-          if (byte2 > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || byte3 > (byte) 0xBF
-              // byte4 trailing-byte test
-              || buffer.get(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-
-      // Finish validation for the sequence.
-      return partialIsValidUtf8(buffer, index, limit);
-    }
-
-    /**
-     * Performs validation for {@link ByteBuffer} instances using the {@link ByteBuffer} API rather
-     * than potentially faster approaches.
-     */
-    private static int partialIsValidUtf8(final ByteBuffer buffer, int index, final int limit) {
-      index += estimateConsecutiveAscii(buffer, index, limit);
-
-      for (;;) {
-        // Optimize for interior runs of ASCII bytes.
-        // TODO(nathanmittler): Consider checking 8 bytes at a time after some threshold?
-        // Maybe after seeing a few in a row that are ASCII, go back to fast mode?
-        int byte1;
-        do {
-          if (index >= limit) {
-            return COMPLETE;
-          }
-        } while ((byte1 = buffer.get(index++)) >= 0);
-
-        // If we're here byte1 is not ASCII. Only need to handle 2-4 byte forms.
-        if (byte1 < (byte) 0xE0) {
-          // Two-byte form (110xxxxx 10xxxxxx)
-          if (index >= limit) {
-            // Incomplete sequence
-            return byte1;
-          }
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2 || buffer.get(index) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-          index++;
-        } else if (byte1 < (byte) 0xF0) {
-          // Three-byte form (1110xxxx 10xxxxxx 10xxxxxx)
-          if (index >= limit - 1) {
-            // Incomplete sequence
-            return incompleteStateFor(buffer, byte1, index, limit - index);
-          }
-
-          final byte byte2 = buffer.get(index++);
-          if (byte2 > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // check for illegal surrogate codepoints
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || buffer.get(index) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-          index++;
-        } else {
-          // Four-byte form (1110xxxx 10xxxxxx 10xxxxxx 10xxxxxx)
-          if (index >= limit - 2) {
-            // Incomplete sequence
-            return incompleteStateFor(buffer, byte1, index, limit - index);
-          }
-
-          // TODO(nathanmittler): Consider using getInt() to improve performance.
-          final int byte2 = buffer.get(index++);
-          if (byte2 > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || buffer.get(index++) > (byte) 0xBF
-              // byte4 trailing-byte test
-              || buffer.get(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-    }
-
-    public boolean isValidUtf8(ByteInput buffer, int index, int limit) {
-      return partialIsValidUtf8(COMPLETE, buffer, index, limit) == COMPLETE;
-    }
-
-    int partialIsValidUtf8(int state, ByteInput bytes, int index, int limit) {
-      if (state != COMPLETE) {
-        // The previous decoding operation was incomplete (or malformed).
-        // We look for a well-formed sequence consisting of bytes from
-        // the previous decoding operation (stored in state) together
-        // with bytes from the array slice.
-        //
-        // We expect such "straddler characters" to be rare.
-
-        if (index >= limit) {  // No bytes? No progress.
-          return state;
-        }
-        int byte1 = (byte) state;
-        // byte1 is never ASCII.
-        if (byte1 < (byte) 0xE0) {
-          // two-byte form
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2
-              // byte2 trailing-byte test
-              || bytes.read(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // three-byte form
-
-          // Get byte2 from saved state or array
-          int byte2 = (byte) ~(state >> 8);
-          if (byte2 == 0) {
-            byte2 = bytes.read(index++);
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          }
-          if (byte2 > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // illegal surrogate codepoint?
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || bytes.read(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // four-byte form
-
-          // Get byte2 and byte3 from saved state or array
-          int byte2 = (byte) ~(state >> 8);
-          int byte3 = 0;
-          if (byte2 == 0) {
-            byte2 = bytes.read(index++);
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          } else {
-            byte3 = (byte) (state >> 16);
-          }
-          if (byte3 == 0) {
-            byte3 = bytes.read(index++);
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2, byte3);
-            }
-          }
-
-          // If we were called with state == MALFORMED, then byte1 is 0xFF,
-          // which never occurs in well-formed UTF-8, and so we will return
-          // MALFORMED again below.
-
-          if (byte2 > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || byte3 > (byte) 0xBF
-              // byte4 trailing-byte test
-              || bytes.read(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-
-      return partialIsValidUtf8(bytes, index, limit);
-    }
-
-    private static int partialIsValidUtf8(ByteInput bytes, int index, int limit) {
-      // Optimize for 100% ASCII (Hotspot loves small simple top-level loops like this).
-      // This simple loop stops when we encounter a byte >= 0x80 (i.e. non-ASCII).
-      while (index < limit && bytes.read(index) >= 0) {
-        index++;
-      }
-
-      return (index >= limit) ? COMPLETE : partialIsValidUtf8NonAscii(bytes, index, limit);
-    }
-
-    private static int partialIsValidUtf8NonAscii(ByteInput bytes, int index, int limit) {
-      for (;;) {
-        int byte1, byte2;
-
-        // Optimize for interior runs of ASCII bytes.
-        do {
-          if (index >= limit) {
-            return COMPLETE;
-          }
-        } while ((byte1 = bytes.read(index++)) >= 0);
-
-        if (byte1 < (byte) 0xE0) {
-          // two-byte form
-
-          if (index >= limit) {
-            // Incomplete sequence
-            return byte1;
-          }
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2
-              || bytes.read(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // three-byte form
-
-          if (index >= limit - 1) { // incomplete sequence
-            return incompleteStateFor(bytes, index, limit);
-          }
-          if ((byte2 = bytes.read(index++)) > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // check for illegal surrogate codepoints
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || bytes.read(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // four-byte form
-
-          if (index >= limit - 2) {  // incomplete sequence
-            return incompleteStateFor(bytes, index, limit);
-          }
-          if ((byte2 = bytes.read(index++)) > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || bytes.read(index++) > (byte) 0xBF
-              // byte4 trailing-byte test
-              || bytes.read(index++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-    }
-
-    /**
-     * Encodes an input character sequence ({@code in}) to UTF-8 in the target array ({@code out}).
-     * For a string, this method is similar to
-     * <pre>{@code
-     * byte[] a = string.getBytes(UTF_8);
-     * System.arraycopy(a, 0, bytes, offset, a.length);
-     * return offset + a.length;
-     * }</pre>
-     *
-     * but is more efficient in both time and space. One key difference is that this method
-     * requires paired surrogates, and therefore does not support chunking.
-     * While {@code String.getBytes(UTF_8)} replaces unpaired surrogates with the default
-     * replacement character, this method throws {@link UnpairedSurrogateException}.
-     *
-     * <p>To ensure sufficient space in the output buffer, either call {@link #encodedLength} to
-     * compute the exact amount needed, or leave room for 
-     * {@code Utf8.MAX_BYTES_PER_CHAR * sequence.length()}, which is the largest possible number
-     * of bytes that any input can be encoded to.
-     *
-     * @param in the input character sequence to be encoded
-     * @param out the target array
-     * @param offset the starting offset in {@code bytes} to start writing at
-     * @param length the length of the {@code bytes}, starting from {@code offset}
-     * @throws UnpairedSurrogateException if {@code sequence} contains ill-formed UTF-16 (unpaired
-     *     surrogates)
-     * @throws ArrayIndexOutOfBoundsException if {@code sequence} encoded in UTF-8 is longer than
-     *     {@code bytes.length - offset}
-     * @return the new offset, equivalent to {@code offset + Utf8.encodedLength(sequence)}
-     */
-    abstract int encodeUtf8(CharSequence in, byte[] out, int offset, int length);
-
-    /**
-     * Encodes an input character sequence ({@code in}) to UTF-8 in the target buffer ({@code out}).
-     * Upon returning from this method, the {@code out} position will point to the position after
-     * the last encoded byte. This method requires paired surrogates, and therefore does not
-     * support chunking.
-     *
-     * <p>To ensure sufficient space in the output buffer, either call {@link #encodedLength} to
-     * compute the exact amount needed, or leave room for
-     * {@code Utf8.MAX_BYTES_PER_CHAR * in.length()}, which is the largest possible number
-     * of bytes that any input can be encoded to.
-     *
-     * @param in the source character sequence to be encoded
-     * @param out the target buffer
-     * @throws UnpairedSurrogateException if {@code in} contains ill-formed UTF-16 (unpaired
-     *     surrogates)
-     * @throws ArrayIndexOutOfBoundsException if {@code in} encoded in UTF-8 is longer than
-     *     {@code out.remaining()}
-     */
-    final void encodeUtf8(CharSequence in, ByteBuffer out) {
-      if (out.hasArray()) {
-        final int offset = out.arrayOffset();
-        int endIndex =
-            Utf8.encode(in, out.array(), offset + out.position(), out.remaining());
-        out.position(endIndex - offset);
-      } else if (out.isDirect()) {
-        encodeUtf8Direct(in, out);
-      } else {
-        encodeUtf8Default(in, out);
-      }
-    }
-
-    /**
-     * Encodes the input character sequence to a direct {@link ByteBuffer} instance.
-     */
-    abstract void encodeUtf8Direct(CharSequence in, ByteBuffer out);
-
-    /**
-     * Encodes the input character sequence to a {@link ByteBuffer} instance using the {@link
-     * ByteBuffer} API, rather than potentially faster approaches.
-     */
-    final void encodeUtf8Default(CharSequence in, ByteBuffer out) {
-      final int inLength = in.length();
-      int outIx = out.position();
-      int inIx = 0;
-
-      // Since ByteBuffer.putXXX() already checks boundaries for us, no need to explicitly check
-      // access. Assume the buffer is big enough and let it handle the out of bounds exception
-      // if it occurs.
-      try {
-        // Designed to take advantage of
-        // https://wikis.oracle.com/display/HotSpotInternals/RangeCheckElimination
-        for (char c; inIx < inLength && (c = in.charAt(inIx)) < 0x80; ++inIx) {
-          out.put(outIx + inIx, (byte) c);
-        }
-        if (inIx == inLength) {
-          // Successfully encoded the entire string.
-          out.position(outIx + inIx);
-          return;
-        }
-
-        outIx += inIx;
-        for (char c; inIx < inLength; ++inIx, ++outIx) {
-          c = in.charAt(inIx);
-          if (c < 0x80) {
-            // One byte (0xxx xxxx)
-            out.put(outIx, (byte) c);
-          } else if (c < 0x800) {
-            // Two bytes (110x xxxx 10xx xxxx)
-
-            // Benchmarks show put performs better than putShort here (for HotSpot).
-            out.put(outIx++, (byte) (0xC0 | (c >>> 6)));
-            out.put(outIx, (byte) (0x80 | (0x3F & c)));
-          } else if (c < MIN_SURROGATE || MAX_SURROGATE < c) {
-            // Three bytes (1110 xxxx 10xx xxxx 10xx xxxx)
-            // Maximum single-char code point is 0xFFFF, 16 bits.
-
-            // Benchmarks show put performs better than putShort here (for HotSpot).
-            out.put(outIx++, (byte) (0xE0 | (c >>> 12)));
-            out.put(outIx++, (byte) (0x80 | (0x3F & (c >>> 6))));
-            out.put(outIx, (byte) (0x80 | (0x3F & c)));
-          } else {
-            // Four bytes (1111 xxxx 10xx xxxx 10xx xxxx 10xx xxxx)
-
-            // Minimum code point represented by a surrogate pair is 0x10000, 17 bits, four UTF-8
-            // bytes
-            final char low;
-            if (inIx + 1 == inLength || !isSurrogatePair(c, (low = in.charAt(++inIx)))) {
-              throw new UnpairedSurrogateException(inIx, inLength);
-            }
-            // TODO(nathanmittler): Consider using putInt() to improve performance.
-            int codePoint = toCodePoint(c, low);
-            out.put(outIx++, (byte) ((0xF << 4) | (codePoint >>> 18)));
-            out.put(outIx++, (byte) (0x80 | (0x3F & (codePoint >>> 12))));
-            out.put(outIx++, (byte) (0x80 | (0x3F & (codePoint >>> 6))));
-            out.put(outIx, (byte) (0x80 | (0x3F & codePoint)));
-          }
-        }
-
-        // Successfully encoded the entire string.
-        out.position(outIx);
-      } catch (IndexOutOfBoundsException e) {
-        // TODO(nathanmittler): Consider making the API throw IndexOutOfBoundsException instead.
-
-        // If we failed in the outer ASCII loop, outIx will not have been updated. In this case,
-        // use inIx to determine the bad write index.
-        int badWriteIndex = out.position() + Math.max(inIx, outIx - out.position() + 1);
-        throw new ArrayIndexOutOfBoundsException(
-            "Failed writing " + in.charAt(inIx) + " at index " + badWriteIndex);
-      }
-    }
-  }
-
-  /**
-   * {@link Processor} implementation that does not use any {@code sun.misc.Unsafe} methods.
-   */
-  static final class SafeProcessor extends Processor {
-    @Override
-    int partialIsValidUtf8(int state, byte[] bytes, int index, int limit) {
-      if (state != COMPLETE) {
-        // The previous decoding operation was incomplete (or malformed).
-        // We look for a well-formed sequence consisting of bytes from
-        // the previous decoding operation (stored in state) together
-        // with bytes from the array slice.
-        //
-        // We expect such "straddler characters" to be rare.
-
-        if (index >= limit) {  // No bytes? No progress.
-          return state;
-        }
-        int byte1 = (byte) state;
-        // byte1 is never ASCII.
-        if (byte1 < (byte) 0xE0) {
-          // two-byte form
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2
-              // byte2 trailing-byte test
-              || bytes[index++] > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // three-byte form
-
-          // Get byte2 from saved state or array
-          int byte2 = (byte) ~(state >> 8);
-          if (byte2 == 0) {
-            byte2 = bytes[index++];
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          }
-          if (byte2 > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // illegal surrogate codepoint?
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || bytes[index++] > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // four-byte form
-
-          // Get byte2 and byte3 from saved state or array
-          int byte2 = (byte) ~(state >> 8);
-          int byte3 = 0;
-          if (byte2 == 0) {
-            byte2 = bytes[index++];
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          } else {
-            byte3 = (byte) (state >> 16);
-          }
-          if (byte3 == 0) {
-            byte3 = bytes[index++];
-            if (index >= limit) {
-              return incompleteStateFor(byte1, byte2, byte3);
-            }
-          }
-
-          // If we were called with state == MALFORMED, then byte1 is 0xFF,
-          // which never occurs in well-formed UTF-8, and so we will return
-          // MALFORMED again below.
-
-          if (byte2 > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || byte3 > (byte) 0xBF
-              // byte4 trailing-byte test
-              || bytes[index++] > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-
-      return partialIsValidUtf8(bytes, index, limit);
-    }
-
-    @Override
-    int partialIsValidUtf8Direct(int state, ByteBuffer buffer, int index, int limit) {
-      // For safe processing, we have to use the ByteBuffer API.
-      return partialIsValidUtf8Default(state, buffer, index, limit);
-    }
-
-    @Override
-    int encodeUtf8(CharSequence in, byte[] out, int offset, int length) {
-      int utf16Length = in.length();
-      int j = offset;
-      int i = 0;
-      int limit = offset + length;
-      // Designed to take advantage of
-      // https://wikis.oracle.com/display/HotSpotInternals/RangeCheckElimination
-      for (char c; i < utf16Length && i + j < limit && (c = in.charAt(i)) < 0x80; i++) {
-        out[j + i] = (byte) c;
-      }
-      if (i == utf16Length) {
-        return j + utf16Length;
-      }
-      j += i;
-      for (char c; i < utf16Length; i++) {
-        c = in.charAt(i);
-        if (c < 0x80 && j < limit) {
-          out[j++] = (byte) c;
-        } else if (c < 0x800 && j <= limit - 2) { // 11 bits, two UTF-8 bytes
-          out[j++] = (byte) ((0xF << 6) | (c >>> 6));
-          out[j++] = (byte) (0x80 | (0x3F & c));
-        } else if ((c < Character.MIN_SURROGATE || Character.MAX_SURROGATE < c) && j <= limit - 3) {
-          // Maximum single-char code point is 0xFFFF, 16 bits, three UTF-8 bytes
-          out[j++] = (byte) ((0xF << 5) | (c >>> 12));
-          out[j++] = (byte) (0x80 | (0x3F & (c >>> 6)));
-          out[j++] = (byte) (0x80 | (0x3F & c));
-        } else if (j <= limit - 4) {
-          // Minimum code point represented by a surrogate pair is 0x10000, 17 bits,
-          // four UTF-8 bytes
-          final char low;
-          if (i + 1 == in.length()
-                  || !Character.isSurrogatePair(c, (low = in.charAt(++i)))) {
-            throw new UnpairedSurrogateException((i - 1), utf16Length);
-          }
-          int codePoint = Character.toCodePoint(c, low);
-          out[j++] = (byte) ((0xF << 4) | (codePoint >>> 18));
-          out[j++] = (byte) (0x80 | (0x3F & (codePoint >>> 12)));
-          out[j++] = (byte) (0x80 | (0x3F & (codePoint >>> 6)));
-          out[j++] = (byte) (0x80 | (0x3F & codePoint));
-        } else {
-          // If we are surrogates and we're not a surrogate pair, always throw an
-          // UnpairedSurrogateException instead of an ArrayOutOfBoundsException.
-          if ((Character.MIN_SURROGATE <= c && c <= Character.MAX_SURROGATE)
-              && (i + 1 == in.length()
-                  || !Character.isSurrogatePair(c, in.charAt(i + 1)))) {
-            throw new UnpairedSurrogateException(i, utf16Length);
-          }
-          throw new ArrayIndexOutOfBoundsException("Failed writing " + c + " at index " + j);
-        }
-      }
-      return j;
-    }
-
-    @Override
-    void encodeUtf8Direct(CharSequence in, ByteBuffer out) {
-      // For safe processing, we have to use the ByteBuffer API.
-      encodeUtf8Default(in, out);
-    }
-
-    private static int partialIsValidUtf8(byte[] bytes, int index, int limit) {
-      // Optimize for 100% ASCII (Hotspot loves small simple top-level loops like this).
-      // This simple loop stops when we encounter a byte >= 0x80 (i.e. non-ASCII).
-      while (index < limit && bytes[index] >= 0) {
-        index++;
-      }
-
-      return (index >= limit) ? COMPLETE : partialIsValidUtf8NonAscii(bytes, index, limit);
-    }
-
-    private static int partialIsValidUtf8NonAscii(byte[] bytes, int index, int limit) {
-      for (;;) {
-        int byte1, byte2;
-
-        // Optimize for interior runs of ASCII bytes.
-        do {
-          if (index >= limit) {
-            return COMPLETE;
-          }
-        } while ((byte1 = bytes[index++]) >= 0);
-
-        if (byte1 < (byte) 0xE0) {
-          // two-byte form
-
-          if (index >= limit) {
-            // Incomplete sequence
-            return byte1;
-          }
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2
-              || bytes[index++] > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // three-byte form
-
-          if (index >= limit - 1) { // incomplete sequence
-            return incompleteStateFor(bytes, index, limit);
-          }
-          if ((byte2 = bytes[index++]) > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // check for illegal surrogate codepoints
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || bytes[index++] > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // four-byte form
-
-          if (index >= limit - 2) {  // incomplete sequence
-            return incompleteStateFor(bytes, index, limit);
-          }
-          if ((byte2 = bytes[index++]) > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || bytes[index++] > (byte) 0xBF
-              // byte4 trailing-byte test
-              || bytes[index++] > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-    }
-  }
-
-  /**
-   * {@link Processor} that uses {@code sun.misc.Unsafe} where possible to improve performance.
-   */
-  static final class UnsafeProcessor extends Processor {
-    /**
-     * Indicates whether or not all required unsafe operations are supported on this platform.
-     */
-    static boolean isAvailable() {
-      return hasUnsafeArrayOperations() && hasUnsafeByteBufferOperations();
-    }
-
-    @Override
-    int partialIsValidUtf8(int state, byte[] bytes, final int index, final int limit) {
-      if ((index | limit | bytes.length - limit) < 0) {
-        throw new ArrayIndexOutOfBoundsException(
-            String.format("Array length=%d, index=%d, limit=%d", bytes.length, index, limit));
-      }
-      long offset = getArrayBaseOffset() + index;
-      final long offsetLimit = getArrayBaseOffset() + limit;
-      if (state != COMPLETE) {
-        // The previous decoding operation was incomplete (or malformed).
-        // We look for a well-formed sequence consisting of bytes from
-        // the previous decoding operation (stored in state) together
-        // with bytes from the array slice.
-        //
-        // We expect such "straddler characters" to be rare.
-
-        if (offset >= offsetLimit) {  // No bytes? No progress.
-          return state;
-        }
-        int byte1 = (byte) state;
-        // byte1 is never ASCII.
-        if (byte1 < (byte) 0xE0) {
-          // two-byte form
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2
-              // byte2 trailing-byte test
-              || UnsafeUtil.getByte(bytes, offset++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // three-byte form
-
-          // Get byte2 from saved state or array
-          int byte2 = (byte) ~(state >> 8);
-          if (byte2 == 0) {
-            byte2 = UnsafeUtil.getByte(bytes, offset++);
-            if (offset >= offsetLimit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          }
-          if (byte2 > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // illegal surrogate codepoint?
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || UnsafeUtil.getByte(bytes, offset++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // four-byte form
-
-          // Get byte2 and byte3 from saved state or array
-          int byte2 = (byte) ~(state >> 8);
-          int byte3 = 0;
-          if (byte2 == 0) {
-            byte2 = UnsafeUtil.getByte(bytes, offset++);
-            if (offset >= offsetLimit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          } else {
-            byte3 = (byte) (state >> 16);
-          }
-          if (byte3 == 0) {
-            byte3 = UnsafeUtil.getByte(bytes, offset++);
-            if (offset >= offsetLimit) {
-              return incompleteStateFor(byte1, byte2, byte3);
-            }
-          }
-
-          // If we were called with state == MALFORMED, then byte1 is 0xFF,
-          // which never occurs in well-formed UTF-8, and so we will return
-          // MALFORMED again below.
-
-          if (byte2 > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || byte3 > (byte) 0xBF
-              // byte4 trailing-byte test
-              || UnsafeUtil.getByte(bytes, offset++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-
-      return partialIsValidUtf8(bytes, offset, (int) (offsetLimit - offset));
-    }
-
-    @Override
-    int partialIsValidUtf8Direct(
-        final int state, ByteBuffer buffer, final int index, final int limit) {
-      if ((index | limit | buffer.limit() - limit) < 0) {
-        throw new ArrayIndexOutOfBoundsException(
-            String.format("buffer limit=%d, index=%d, limit=%d", buffer.limit(), index, limit));
-      }
-      long address = addressOffset(buffer) + index;
-      final long addressLimit = address + (limit - index);
-      if (state != COMPLETE) {
-        // The previous decoding operation was incomplete (or malformed).
-        // We look for a well-formed sequence consisting of bytes from
-        // the previous decoding operation (stored in state) together
-        // with bytes from the array slice.
-        //
-        // We expect such "straddler characters" to be rare.
-
-        if (address >= addressLimit) { // No bytes? No progress.
-          return state;
-        }
-
-        final int byte1 = (byte) state;
-        // byte1 is never ASCII.
-        if (byte1 < (byte) 0xE0) {
-          // two-byte form
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2
-              // byte2 trailing-byte test
-              || UnsafeUtil.getByte(address++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // three-byte form
-
-          // Get byte2 from saved state or array
-          int byte2 = (byte) ~(state >> 8);
-          if (byte2 == 0) {
-            byte2 = UnsafeUtil.getByte(address++);
-            if (address >= addressLimit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          }
-          if (byte2 > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // illegal surrogate codepoint?
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || UnsafeUtil.getByte(address++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // four-byte form
-
-          // Get byte2 and byte3 from saved state or array
-          int byte2 = (byte) ~(state >> 8);
-          int byte3 = 0;
-          if (byte2 == 0) {
-            byte2 = UnsafeUtil.getByte(address++);
-            if (address >= addressLimit) {
-              return incompleteStateFor(byte1, byte2);
-            }
-          } else {
-            byte3 = (byte) (state >> 16);
-          }
-          if (byte3 == 0) {
-            byte3 = UnsafeUtil.getByte(address++);
-            if (address >= addressLimit) {
-              return incompleteStateFor(byte1, byte2, byte3);
-            }
-          }
-
-          // If we were called with state == MALFORMED, then byte1 is 0xFF,
-          // which never occurs in well-formed UTF-8, and so we will return
-          // MALFORMED again below.
-
-          if (byte2 > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || byte3 > (byte) 0xBF
-              // byte4 trailing-byte test
-              || UnsafeUtil.getByte(address++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-
-      return partialIsValidUtf8(address, (int) (addressLimit - address));
-    }
-
-    @Override
-    int encodeUtf8(final CharSequence in, final byte[] out, final int offset, final int length) {
-      long outIx = getArrayBaseOffset() + offset;
-      final long outLimit = outIx + length;
-      final int inLimit = in.length();
-      if (inLimit > length || out.length - length < offset) {
-        // Not even enough room for an ASCII-encoded string.
-        throw new ArrayIndexOutOfBoundsException(
-            "Failed writing " + in.charAt(inLimit - 1) + " at index " + (offset + length));
-      }
-
-      // Designed to take advantage of
-      // https://wikis.oracle.com/display/HotSpotInternals/RangeCheckElimination
-      int inIx = 0;
-      for (char c; inIx < inLimit && (c = in.charAt(inIx)) < 0x80; ++inIx) {
-        UnsafeUtil.putByte(out, outIx++, (byte) c);
-      }
-      if (inIx == inLimit) {
-        // We're done, it was ASCII encoded.
-        return (int) (outIx - getArrayBaseOffset());
-      }
-
-      for (char c; inIx < inLimit; ++inIx) {
-        c = in.charAt(inIx);
-        if (c < 0x80 && outIx < outLimit) {
-          UnsafeUtil.putByte(out, outIx++, (byte) c);
-        } else if (c < 0x800 && outIx <= outLimit - 2L) { // 11 bits, two UTF-8 bytes
-          UnsafeUtil.putByte(out, outIx++, (byte) ((0xF << 6) | (c >>> 6)));
-          UnsafeUtil.putByte(out, outIx++, (byte) (0x80 | (0x3F & c)));
-        } else if ((c < MIN_SURROGATE || MAX_SURROGATE < c) && outIx <= outLimit - 3L) {
-          // Maximum single-char code point is 0xFFFF, 16 bits, three UTF-8 bytes
-          UnsafeUtil.putByte(out, outIx++, (byte) ((0xF << 5) | (c >>> 12)));
-          UnsafeUtil.putByte(out, outIx++, (byte) (0x80 | (0x3F & (c >>> 6))));
-          UnsafeUtil.putByte(out, outIx++, (byte) (0x80 | (0x3F & c)));
-        } else if (outIx <= outLimit - 4L) {
-          // Minimum code point represented by a surrogate pair is 0x10000, 17 bits, four UTF-8
-          // bytes
-          final char low;
-          if (inIx + 1 == inLimit || !isSurrogatePair(c, (low = in.charAt(++inIx)))) {
-            throw new UnpairedSurrogateException((inIx - 1), inLimit);
-          }
-          int codePoint = toCodePoint(c, low);
-          UnsafeUtil.putByte(out, outIx++, (byte) ((0xF << 4) | (codePoint >>> 18)));
-          UnsafeUtil.putByte(out, outIx++, (byte) (0x80 | (0x3F & (codePoint >>> 12))));
-          UnsafeUtil.putByte(out, outIx++, (byte) (0x80 | (0x3F & (codePoint >>> 6))));
-          UnsafeUtil.putByte(out, outIx++, (byte) (0x80 | (0x3F & codePoint)));
-        } else {
-          if ((MIN_SURROGATE <= c && c <= MAX_SURROGATE)
-              && (inIx + 1 == inLimit || !isSurrogatePair(c, in.charAt(inIx + 1)))) {
-            // We are surrogates and we're not a surrogate pair.
-            throw new UnpairedSurrogateException(inIx, inLimit);
-          }
-          // Not enough space in the output buffer.
-          throw new ArrayIndexOutOfBoundsException("Failed writing " + c + " at index " + outIx);
-        }
-      }
-
-      // All bytes have been encoded.
-      return (int) (outIx - getArrayBaseOffset());
-    }
-
-    @Override
-    void encodeUtf8Direct(CharSequence in, ByteBuffer out) {
-      final long address = addressOffset(out);
-      long outIx = address + out.position();
-      final long outLimit = address + out.limit();
-      final int inLimit = in.length();
-      if (inLimit > outLimit - outIx) {
-        // Not even enough room for an ASCII-encoded string.
-        throw new ArrayIndexOutOfBoundsException(
-            "Failed writing " + in.charAt(inLimit - 1) + " at index " + out.limit());
-      }
-
-      // Designed to take advantage of
-      // https://wikis.oracle.com/display/HotSpotInternals/RangeCheckElimination
-      int inIx = 0;
-      for (char c; inIx < inLimit && (c = in.charAt(inIx)) < 0x80; ++inIx) {
-        UnsafeUtil.putByte(outIx++, (byte) c);
-      }
-      if (inIx == inLimit) {
-        // We're done, it was ASCII encoded.
-        out.position((int) (outIx - address));
-        return;
-      }
-
-      for (char c; inIx < inLimit; ++inIx) {
-        c = in.charAt(inIx);
-        if (c < 0x80 && outIx < outLimit) {
-          UnsafeUtil.putByte(outIx++, (byte) c);
-        } else if (c < 0x800 && outIx <= outLimit - 2L) { // 11 bits, two UTF-8 bytes
-          UnsafeUtil.putByte(outIx++, (byte) ((0xF << 6) | (c >>> 6)));
-          UnsafeUtil.putByte(outIx++, (byte) (0x80 | (0x3F & c)));
-        } else if ((c < MIN_SURROGATE || MAX_SURROGATE < c) && outIx <= outLimit - 3L) {
-          // Maximum single-char code point is 0xFFFF, 16 bits, three UTF-8 bytes
-          UnsafeUtil.putByte(outIx++, (byte) ((0xF << 5) | (c >>> 12)));
-          UnsafeUtil.putByte(outIx++, (byte) (0x80 | (0x3F & (c >>> 6))));
-          UnsafeUtil.putByte(outIx++, (byte) (0x80 | (0x3F & c)));
-        } else if (outIx <= outLimit - 4L) {
-          // Minimum code point represented by a surrogate pair is 0x10000, 17 bits, four UTF-8
-          // bytes
-          final char low;
-          if (inIx + 1 == inLimit || !isSurrogatePair(c, (low = in.charAt(++inIx)))) {
-            throw new UnpairedSurrogateException((inIx - 1), inLimit);
-          }
-          int codePoint = toCodePoint(c, low);
-          UnsafeUtil.putByte(outIx++, (byte) ((0xF << 4) | (codePoint >>> 18)));
-          UnsafeUtil.putByte(outIx++, (byte) (0x80 | (0x3F & (codePoint >>> 12))));
-          UnsafeUtil.putByte(outIx++, (byte) (0x80 | (0x3F & (codePoint >>> 6))));
-          UnsafeUtil.putByte(outIx++, (byte) (0x80 | (0x3F & codePoint)));
-        } else {
-          if ((MIN_SURROGATE <= c && c <= MAX_SURROGATE)
-              && (inIx + 1 == inLimit || !isSurrogatePair(c, in.charAt(inIx + 1)))) {
-            // We are surrogates and we're not a surrogate pair.
-            throw new UnpairedSurrogateException(inIx, inLimit);
-          }
-          // Not enough space in the output buffer.
-          throw new ArrayIndexOutOfBoundsException("Failed writing " + c + " at index " + outIx);
-        }
-      }
-
-      // All bytes have been encoded.
-      out.position((int) (outIx - address));
-    }
-
-    /**
-     * Counts (approximately) the number of consecutive ASCII characters starting from the given
-     * position, using the most efficient method available to the platform.
-     *
-     * @param bytes the array containing the character sequence
-     * @param offset the offset position of the index (same as index + arrayBaseOffset)
-     * @param maxChars the maximum number of characters to count
-     * @return the number of ASCII characters found. The stopping position will be at or
-     * before the first non-ASCII byte.
-     */
-    private static int unsafeEstimateConsecutiveAscii(
-        byte[] bytes, long offset, final int maxChars) {
-      int remaining = maxChars;
-      if (remaining < UNSAFE_COUNT_ASCII_THRESHOLD) {
-        // Don't bother with small strings.
-        return 0;
-      }
-
-      // Read bytes until 8-byte aligned so that we can read longs in the loop below.
-      // Byte arrays are already either 8 or 16-byte aligned, so we just need to make sure that
-      // the index (relative to the start of the array) is also 8-byte aligned. We do this by
-      // ANDing the index with 7 to determine the number of bytes that need to be read before
-      // we're 8-byte aligned.
-      final int unaligned = (int) offset & 7;
-      for (int j = unaligned; j > 0; j--) {
-        if (UnsafeUtil.getByte(bytes, offset++) < 0) {
-          return unaligned - j;
-        }
-      }
-
-      // This simple loop stops when we encounter a byte >= 0x80 (i.e. non-ASCII).
-      // To speed things up further, we're reading longs instead of bytes so we use a mask to
-      // determine if any byte in the current long is non-ASCII.
-      remaining -= unaligned;
-      for (; remaining >= 8 && (UnsafeUtil.getLong(bytes, offset) & ASCII_MASK_LONG) == 0;
-          offset += 8, remaining -= 8) {}
-      return maxChars - remaining;
-    }
-
-    /**
-     * Same as {@link Utf8#estimateConsecutiveAscii(ByteBuffer, int, int)} except that it uses the
-     * most efficient method available to the platform.
-     */
-    private static int unsafeEstimateConsecutiveAscii(long address, final int maxChars) {
-      int remaining = maxChars;
-      if (remaining < UNSAFE_COUNT_ASCII_THRESHOLD) {
-        // Don't bother with small strings.
-        return 0;
-      }
-
-      // Read bytes until 8-byte aligned so that we can read longs in the loop below.
-      // We do this by ANDing the address with 7 to determine the number of bytes that need to
-      // be read before we're 8-byte aligned.
-      final int unaligned = (int) address & 7;
-      for (int j = unaligned; j > 0; j--) {
-        if (UnsafeUtil.getByte(address++) < 0) {
-          return unaligned - j;
-        }
-      }
-
-      // This simple loop stops when we encounter a byte >= 0x80 (i.e. non-ASCII).
-      // To speed things up further, we're reading longs instead of bytes so we use a mask to
-      // determine if any byte in the current long is non-ASCII.
-      remaining -= unaligned;
-      for (; remaining >= 8 && (UnsafeUtil.getLong(address) & ASCII_MASK_LONG) == 0;
-          address += 8, remaining -= 8) {}
-      return maxChars - remaining;
-    }
-
-    private static int partialIsValidUtf8(final byte[] bytes, long offset, int remaining) {
-      // Skip past ASCII characters as quickly as possible. 
-      final int skipped = unsafeEstimateConsecutiveAscii(bytes, offset, remaining);
-      remaining -= skipped;
-      offset += skipped;
-
-      for (;;) {
-        // Optimize for interior runs of ASCII bytes.
-        // TODO(nathanmittler): Consider checking 8 bytes at a time after some threshold?
-        // Maybe after seeing a few in a row that are ASCII, go back to fast mode?
-        int byte1 = 0;
-        for (; remaining > 0 && (byte1 = UnsafeUtil.getByte(bytes, offset++)) >= 0; --remaining) {
-        }
-        if (remaining == 0) {
-          return COMPLETE;
-        }
-        remaining--;
-
-        // If we're here byte1 is not ASCII. Only need to handle 2-4 byte forms.
-        if (byte1 < (byte) 0xE0) {
-          // Two-byte form (110xxxxx 10xxxxxx)
-          if (remaining == 0) {
-            // Incomplete sequence
-            return byte1;
-          }
-          remaining--;
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2
-              || UnsafeUtil.getByte(bytes, offset++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // Three-byte form (1110xxxx 10xxxxxx 10xxxxxx)
-          if (remaining < 2) {
-            // Incomplete sequence
-            return unsafeIncompleteStateFor(bytes, byte1, offset, remaining);
-          }
-          remaining -= 2;
-
-          final int byte2;
-          if ((byte2 = UnsafeUtil.getByte(bytes, offset++)) > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // check for illegal surrogate codepoints
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || UnsafeUtil.getByte(bytes, offset++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // Four-byte form (1110xxxx 10xxxxxx 10xxxxxx 10xxxxxx)
-          if (remaining < 3) {
-            // Incomplete sequence
-            return unsafeIncompleteStateFor(bytes, byte1, offset, remaining);
-          }
-          remaining -= 3;
-
-          final int byte2;
-          if ((byte2 = UnsafeUtil.getByte(bytes, offset++)) > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || UnsafeUtil.getByte(bytes, offset++) > (byte) 0xBF
-              // byte4 trailing-byte test
-              || UnsafeUtil.getByte(bytes, offset++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-    }
-
-    private static int partialIsValidUtf8(long address, int remaining) {
-      // Skip past ASCII characters as quickly as possible.
-      final int skipped = unsafeEstimateConsecutiveAscii(address, remaining);
-      address += skipped;
-      remaining -= skipped;
-
-      for (;;) {
-        // Optimize for interior runs of ASCII bytes.
-        // TODO(nathanmittler): Consider checking 8 bytes at a time after some threshold?
-        // Maybe after seeing a few in a row that are ASCII, go back to fast mode?
-        int byte1 = 0;
-        for (; remaining > 0 && (byte1 = UnsafeUtil.getByte(address++)) >= 0; --remaining) {
-        }
-        if (remaining == 0) {
-          return COMPLETE;
-        }
-        remaining--;
-
-        if (byte1 < (byte) 0xE0) {
-          // Two-byte form
-
-          if (remaining == 0) {
-            // Incomplete sequence
-            return byte1;
-          }
-          remaining--;
-
-          // Simultaneously checks for illegal trailing-byte in
-          // leading position and overlong 2-byte form.
-          if (byte1 < (byte) 0xC2 || UnsafeUtil.getByte(address++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else if (byte1 < (byte) 0xF0) {
-          // Three-byte form
-
-          if (remaining < 2) {
-            // Incomplete sequence
-            return unsafeIncompleteStateFor(address, byte1, remaining);
-          }
-          remaining -= 2;
-
-          final byte byte2 = UnsafeUtil.getByte(address++);
-          if (byte2 > (byte) 0xBF
-              // overlong? 5 most significant bits must not all be zero
-              || (byte1 == (byte) 0xE0 && byte2 < (byte) 0xA0)
-              // check for illegal surrogate codepoints
-              || (byte1 == (byte) 0xED && byte2 >= (byte) 0xA0)
-              // byte3 trailing-byte test
-              || UnsafeUtil.getByte(address++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        } else {
-          // Four-byte form
-
-          if (remaining < 3) {
-            // Incomplete sequence
-            return unsafeIncompleteStateFor(address, byte1, remaining);
-          }
-          remaining -= 3;
-
-          final byte byte2 = UnsafeUtil.getByte(address++);
-          if (byte2 > (byte) 0xBF
-              // Check that 1 <= plane <= 16.  Tricky optimized form of:
-              // if (byte1 > (byte) 0xF4 ||
-              //     byte1 == (byte) 0xF0 && byte2 < (byte) 0x90 ||
-              //     byte1 == (byte) 0xF4 && byte2 > (byte) 0x8F)
-              || (((byte1 << 28) + (byte2 - (byte) 0x90)) >> 30) != 0
-              // byte3 trailing-byte test
-              || UnsafeUtil.getByte(address++) > (byte) 0xBF
-              // byte4 trailing-byte test
-              || UnsafeUtil.getByte(address++) > (byte) 0xBF) {
-            return MALFORMED;
-          }
-        }
-      }
-    }
-
-    private static int unsafeIncompleteStateFor(byte[] bytes, int byte1, long offset,
-        int remaining) {
-      switch (remaining) {
-        case 0: {
-          return incompleteStateFor(byte1);
-        }
-        case 1: {
-          return incompleteStateFor(byte1, UnsafeUtil.getByte(bytes, offset));
-        }
-        case 2: {
-          return incompleteStateFor(byte1, UnsafeUtil.getByte(bytes, offset),
-              UnsafeUtil.getByte(bytes, offset + 1));
-        }
-        default: {
-          throw new AssertionError();
-        }
-      }
-    }
-
-    private static int unsafeIncompleteStateFor(long address, final int byte1, int remaining) {
-      switch (remaining) {
-        case 0: {
-          return incompleteStateFor(byte1);
-        }
-        case 1: {
-          return incompleteStateFor(byte1, UnsafeUtil.getByte(address));
-        }
-        case 2: {
-          return incompleteStateFor(byte1, UnsafeUtil.getByte(address),
-              UnsafeUtil.getByte(address + 1));
-        }
-        default: {
-          throw new AssertionError();
-        }
-      }
-    }
-  }
-
-  private Utf8() {}
-}


[52/52] hbase git commit: HBASE-17056 Remove checked in PB generated files; ADDENDUM

Posted by st...@apache.org.
HBASE-17056 Remove checked in PB generated files; ADDENDUM


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/6440509d
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/6440509d
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/6440509d

Branch: refs/heads/branch-2
Commit: 6440509d7d28b0a0ecf027352d64605d7085ff29
Parents: ee70b1d
Author: Michael Stack <st...@apache.org>
Authored: Wed Aug 2 09:54:03 2017 -0700
Committer: Michael Stack <st...@apache.org>
Committed: Wed Aug 2 09:54:03 2017 -0700

----------------------------------------------------------------------
 hbase-spark/pom.xml | 1 -
 pom.xml             | 5 -----
 2 files changed, 6 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/6440509d/hbase-spark/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-spark/pom.xml b/hbase-spark/pom.xml
index 8138a5d..806b024 100644
--- a/hbase-spark/pom.xml
+++ b/hbase-spark/pom.xml
@@ -184,7 +184,6 @@
           <artifactId>netty</artifactId>
         </exclusion>
         <exclusion>
->>>>>>> 7a6de1bd42... HBASE-17056 Remove checked in PB generated files
           <groupId>com.google.code.findbugs</groupId>
           <artifactId>jsr305</artifactId>
         </exclusion>

http://git-wip-us.apache.org/repos/asf/hbase/blob/6440509d/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 53092d1..2bd43b9 100644
--- a/pom.xml
+++ b/pom.xml
@@ -1689,11 +1689,6 @@
         <version>${metrics-core.version}</version>
       </dependency>
       <dependency>
-        <groupId>org.apache.hbase.thirdparty</groupId>
-        <artifactId>hbase-shaded-miscellaneous</artifactId>
-        <version>${hbase-thirdparty.version}</version>
-      </dependency>
-      <dependency>
         <groupId>commons-collections</groupId>
         <artifactId>commons-collections</artifactId>
         <version>${collections.version}</version>


[06/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java
deleted file mode 100644
index 46e9c01..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/ipc/protobuf/generated/TestProcedureProtos.java
+++ /dev/null
@@ -1,608 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: TestProcedure.proto
-
-package org.apache.hadoop.hbase.shaded.ipc.protobuf.generated;
-
-public final class TestProcedureProtos {
-  private TestProcedureProtos() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  public interface TestTableDDLStateDataOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:TestTableDDLStateData)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>required string table_name = 1;</code>
-     */
-    boolean hasTableName();
-    /**
-     * <code>required string table_name = 1;</code>
-     */
-    java.lang.String getTableName();
-    /**
-     * <code>required string table_name = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getTableNameBytes();
-  }
-  /**
-   * Protobuf type {@code TestTableDDLStateData}
-   */
-  public  static final class TestTableDDLStateData extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:TestTableDDLStateData)
-      TestTableDDLStateDataOrBuilder {
-    // Use TestTableDDLStateData.newBuilder() to construct.
-    private TestTableDDLStateData(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private TestTableDDLStateData() {
-      tableName_ = "";
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private TestTableDDLStateData(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000001;
-              tableName_ = bs;
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int TABLE_NAME_FIELD_NUMBER = 1;
-    private volatile java.lang.Object tableName_;
-    /**
-     * <code>required string table_name = 1;</code>
-     */
-    public boolean hasTableName() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required string table_name = 1;</code>
-     */
-    public java.lang.String getTableName() {
-      java.lang.Object ref = tableName_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          tableName_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <code>required string table_name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getTableNameBytes() {
-      java.lang.Object ref = tableName_;
-      if (ref instanceof java.lang.String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        tableName_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      if (!hasTableName()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, tableName_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, tableName_);
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData other = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData) obj;
-
-      boolean result = true;
-      result = result && (hasTableName() == other.hasTableName());
-      if (hasTableName()) {
-        result = result && getTableName()
-            .equals(other.getTableName());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasTableName()) {
-        hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
-        hash = (53 * hash) + getTableName().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code TestTableDDLStateData}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:TestTableDDLStateData)
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateDataOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.class, org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        tableName_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.internal_static_TestTableDDLStateData_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData build() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData buildPartial() {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData result = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.tableName_ = tableName_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData other) {
-        if (other == org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData.getDefaultInstance()) return this;
-        if (other.hasTableName()) {
-          bitField0_ |= 0x00000001;
-          tableName_ = other.tableName_;
-          onChanged();
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasTableName()) {
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private java.lang.Object tableName_ = "";
-      /**
-       * <code>required string table_name = 1;</code>
-       */
-      public boolean hasTableName() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required string table_name = 1;</code>
-       */
-      public java.lang.String getTableName() {
-        java.lang.Object ref = tableName_;
-        if (!(ref instanceof java.lang.String)) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-              (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-          java.lang.String s = bs.toStringUtf8();
-          if (bs.isValidUtf8()) {
-            tableName_ = s;
-          }
-          return s;
-        } else {
-          return (java.lang.String) ref;
-        }
-      }
-      /**
-       * <code>required string table_name = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-          getTableNameBytes() {
-        java.lang.Object ref = tableName_;
-        if (ref instanceof String) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                  (java.lang.String) ref);
-          tableName_ = b;
-          return b;
-        } else {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        }
-      }
-      /**
-       * <code>required string table_name = 1;</code>
-       */
-      public Builder setTableName(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        tableName_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string table_name = 1;</code>
-       */
-      public Builder clearTableName() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        tableName_ = getDefaultInstance().getTableName();
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string table_name = 1;</code>
-       */
-      public Builder setTableNameBytes(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        tableName_ = value;
-        onChanged();
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:TestTableDDLStateData)
-    }
-
-    // @@protoc_insertion_point(class_scope:TestTableDDLStateData)
-    private static final org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TestTableDDLStateData>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<TestTableDDLStateData>() {
-      public TestTableDDLStateData parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new TestTableDDLStateData(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TestTableDDLStateData> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<TestTableDDLStateData> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.ipc.protobuf.generated.TestProcedureProtos.TestTableDDLStateData getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_TestTableDDLStateData_descriptor;
-  private static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_TestTableDDLStateData_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\023TestProcedure.proto\"+\n\025TestTableDDLSta" +
-      "teData\022\022\n\ntable_name\030\001 \002(\tBO\n5org.apache" +
-      ".hadoop.hbase.shaded.ipc.protobuf.genera" +
-      "tedB\023TestProcedureProtos\210\001\001"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_TestTableDDLStateData_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_TestTableDDLStateData_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_TestTableDDLStateData_descriptor,
-        new java.lang.String[] { "TableName", });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[07/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/compiler/PluginProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/compiler/PluginProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/compiler/PluginProtos.java
deleted file mode 100644
index 71975c2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/compiler/PluginProtos.java
+++ /dev/null
@@ -1,5424 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/compiler/plugin.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler;
-
-public final class PluginProtos {
-  private PluginProtos() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  public interface VersionOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:google.protobuf.compiler.Version)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <code>optional int32 major = 1;</code>
-     */
-    boolean hasMajor();
-    /**
-     * <code>optional int32 major = 1;</code>
-     */
-    int getMajor();
-
-    /**
-     * <code>optional int32 minor = 2;</code>
-     */
-    boolean hasMinor();
-    /**
-     * <code>optional int32 minor = 2;</code>
-     */
-    int getMinor();
-
-    /**
-     * <code>optional int32 patch = 3;</code>
-     */
-    boolean hasPatch();
-    /**
-     * <code>optional int32 patch = 3;</code>
-     */
-    int getPatch();
-
-    /**
-     * <pre>
-     * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-     * be empty for mainline stable releases.
-     * </pre>
-     *
-     * <code>optional string suffix = 4;</code>
-     */
-    boolean hasSuffix();
-    /**
-     * <pre>
-     * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-     * be empty for mainline stable releases.
-     * </pre>
-     *
-     * <code>optional string suffix = 4;</code>
-     */
-    java.lang.String getSuffix();
-    /**
-     * <pre>
-     * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-     * be empty for mainline stable releases.
-     * </pre>
-     *
-     * <code>optional string suffix = 4;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getSuffixBytes();
-  }
-  /**
-   * <pre>
-   * The version number of protocol compiler.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.compiler.Version}
-   */
-  public  static final class Version extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:google.protobuf.compiler.Version)
-      VersionOrBuilder {
-    // Use Version.newBuilder() to construct.
-    private Version(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private Version() {
-      major_ = 0;
-      minor_ = 0;
-      patch_ = 0;
-      suffix_ = "";
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private Version(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              major_ = input.readInt32();
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              minor_ = input.readInt32();
-              break;
-            }
-            case 24: {
-              bitField0_ |= 0x00000004;
-              patch_ = input.readInt32();
-              break;
-            }
-            case 34: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000008;
-              suffix_ = bs;
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_Version_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_Version_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int MAJOR_FIELD_NUMBER = 1;
-    private int major_;
-    /**
-     * <code>optional int32 major = 1;</code>
-     */
-    public boolean hasMajor() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>optional int32 major = 1;</code>
-     */
-    public int getMajor() {
-      return major_;
-    }
-
-    public static final int MINOR_FIELD_NUMBER = 2;
-    private int minor_;
-    /**
-     * <code>optional int32 minor = 2;</code>
-     */
-    public boolean hasMinor() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional int32 minor = 2;</code>
-     */
-    public int getMinor() {
-      return minor_;
-    }
-
-    public static final int PATCH_FIELD_NUMBER = 3;
-    private int patch_;
-    /**
-     * <code>optional int32 patch = 3;</code>
-     */
-    public boolean hasPatch() {
-      return ((bitField0_ & 0x00000004) == 0x00000004);
-    }
-    /**
-     * <code>optional int32 patch = 3;</code>
-     */
-    public int getPatch() {
-      return patch_;
-    }
-
-    public static final int SUFFIX_FIELD_NUMBER = 4;
-    private volatile java.lang.Object suffix_;
-    /**
-     * <pre>
-     * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-     * be empty for mainline stable releases.
-     * </pre>
-     *
-     * <code>optional string suffix = 4;</code>
-     */
-    public boolean hasSuffix() {
-      return ((bitField0_ & 0x00000008) == 0x00000008);
-    }
-    /**
-     * <pre>
-     * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-     * be empty for mainline stable releases.
-     * </pre>
-     *
-     * <code>optional string suffix = 4;</code>
-     */
-    public java.lang.String getSuffix() {
-      java.lang.Object ref = suffix_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          suffix_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <pre>
-     * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-     * be empty for mainline stable releases.
-     * </pre>
-     *
-     * <code>optional string suffix = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getSuffixBytes() {
-      java.lang.Object ref = suffix_;
-      if (ref instanceof java.lang.String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        suffix_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeInt32(1, major_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeInt32(2, minor_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        output.writeInt32(3, patch_);
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, suffix_);
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeInt32Size(1, major_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeInt32Size(2, minor_);
-      }
-      if (((bitField0_ & 0x00000004) == 0x00000004)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeInt32Size(3, patch_);
-      }
-      if (((bitField0_ & 0x00000008) == 0x00000008)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, suffix_);
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version) obj;
-
-      boolean result = true;
-      result = result && (hasMajor() == other.hasMajor());
-      if (hasMajor()) {
-        result = result && (getMajor()
-            == other.getMajor());
-      }
-      result = result && (hasMinor() == other.hasMinor());
-      if (hasMinor()) {
-        result = result && (getMinor()
-            == other.getMinor());
-      }
-      result = result && (hasPatch() == other.hasPatch());
-      if (hasPatch()) {
-        result = result && (getPatch()
-            == other.getPatch());
-      }
-      result = result && (hasSuffix() == other.hasSuffix());
-      if (hasSuffix()) {
-        result = result && getSuffix()
-            .equals(other.getSuffix());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (hasMajor()) {
-        hash = (37 * hash) + MAJOR_FIELD_NUMBER;
-        hash = (53 * hash) + getMajor();
-      }
-      if (hasMinor()) {
-        hash = (37 * hash) + MINOR_FIELD_NUMBER;
-        hash = (53 * hash) + getMinor();
-      }
-      if (hasPatch()) {
-        hash = (37 * hash) + PATCH_FIELD_NUMBER;
-        hash = (53 * hash) + getPatch();
-      }
-      if (hasSuffix()) {
-        hash = (37 * hash) + SUFFIX_FIELD_NUMBER;
-        hash = (53 * hash) + getSuffix().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * <pre>
-     * The version number of protocol compiler.
-     * </pre>
-     *
-     * Protobuf type {@code google.protobuf.compiler.Version}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:google.protobuf.compiler.Version)
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.VersionOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_Version_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_Version_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        major_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        minor_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        patch_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000004);
-        suffix_ = "";
-        bitField0_ = (bitField0_ & ~0x00000008);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_Version_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version build() {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version buildPartial() {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.major_ = major_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.minor_ = minor_;
-        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
-          to_bitField0_ |= 0x00000004;
-        }
-        result.patch_ = patch_;
-        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
-          to_bitField0_ |= 0x00000008;
-        }
-        result.suffix_ = suffix_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version other) {
-        if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.getDefaultInstance()) return this;
-        if (other.hasMajor()) {
-          setMajor(other.getMajor());
-        }
-        if (other.hasMinor()) {
-          setMinor(other.getMinor());
-        }
-        if (other.hasPatch()) {
-          setPatch(other.getPatch());
-        }
-        if (other.hasSuffix()) {
-          bitField0_ |= 0x00000008;
-          suffix_ = other.suffix_;
-          onChanged();
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private int major_ ;
-      /**
-       * <code>optional int32 major = 1;</code>
-       */
-      public boolean hasMajor() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>optional int32 major = 1;</code>
-       */
-      public int getMajor() {
-        return major_;
-      }
-      /**
-       * <code>optional int32 major = 1;</code>
-       */
-      public Builder setMajor(int value) {
-        bitField0_ |= 0x00000001;
-        major_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional int32 major = 1;</code>
-       */
-      public Builder clearMajor() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        major_ = 0;
-        onChanged();
-        return this;
-      }
-
-      private int minor_ ;
-      /**
-       * <code>optional int32 minor = 2;</code>
-       */
-      public boolean hasMinor() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional int32 minor = 2;</code>
-       */
-      public int getMinor() {
-        return minor_;
-      }
-      /**
-       * <code>optional int32 minor = 2;</code>
-       */
-      public Builder setMinor(int value) {
-        bitField0_ |= 0x00000002;
-        minor_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional int32 minor = 2;</code>
-       */
-      public Builder clearMinor() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        minor_ = 0;
-        onChanged();
-        return this;
-      }
-
-      private int patch_ ;
-      /**
-       * <code>optional int32 patch = 3;</code>
-       */
-      public boolean hasPatch() {
-        return ((bitField0_ & 0x00000004) == 0x00000004);
-      }
-      /**
-       * <code>optional int32 patch = 3;</code>
-       */
-      public int getPatch() {
-        return patch_;
-      }
-      /**
-       * <code>optional int32 patch = 3;</code>
-       */
-      public Builder setPatch(int value) {
-        bitField0_ |= 0x00000004;
-        patch_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional int32 patch = 3;</code>
-       */
-      public Builder clearPatch() {
-        bitField0_ = (bitField0_ & ~0x00000004);
-        patch_ = 0;
-        onChanged();
-        return this;
-      }
-
-      private java.lang.Object suffix_ = "";
-      /**
-       * <pre>
-       * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-       * be empty for mainline stable releases.
-       * </pre>
-       *
-       * <code>optional string suffix = 4;</code>
-       */
-      public boolean hasSuffix() {
-        return ((bitField0_ & 0x00000008) == 0x00000008);
-      }
-      /**
-       * <pre>
-       * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-       * be empty for mainline stable releases.
-       * </pre>
-       *
-       * <code>optional string suffix = 4;</code>
-       */
-      public java.lang.String getSuffix() {
-        java.lang.Object ref = suffix_;
-        if (!(ref instanceof java.lang.String)) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-              (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-          java.lang.String s = bs.toStringUtf8();
-          if (bs.isValidUtf8()) {
-            suffix_ = s;
-          }
-          return s;
-        } else {
-          return (java.lang.String) ref;
-        }
-      }
-      /**
-       * <pre>
-       * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-       * be empty for mainline stable releases.
-       * </pre>
-       *
-       * <code>optional string suffix = 4;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-          getSuffixBytes() {
-        java.lang.Object ref = suffix_;
-        if (ref instanceof String) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                  (java.lang.String) ref);
-          suffix_ = b;
-          return b;
-        } else {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        }
-      }
-      /**
-       * <pre>
-       * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-       * be empty for mainline stable releases.
-       * </pre>
-       *
-       * <code>optional string suffix = 4;</code>
-       */
-      public Builder setSuffix(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000008;
-        suffix_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-       * be empty for mainline stable releases.
-       * </pre>
-       *
-       * <code>optional string suffix = 4;</code>
-       */
-      public Builder clearSuffix() {
-        bitField0_ = (bitField0_ & ~0x00000008);
-        suffix_ = getDefaultInstance().getSuffix();
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * A suffix for alpha, beta or rc release, e.g., "alpha-1", "rc2". It should
-       * be empty for mainline stable releases.
-       * </pre>
-       *
-       * <code>optional string suffix = 4;</code>
-       */
-      public Builder setSuffixBytes(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000008;
-        suffix_ = value;
-        onChanged();
-        return this;
-      }
-      public final Builder setUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.setUnknownFields(unknownFields);
-      }
-
-      public final Builder mergeUnknownFields(
-          final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-        return super.mergeUnknownFields(unknownFields);
-      }
-
-
-      // @@protoc_insertion_point(builder_scope:google.protobuf.compiler.Version)
-    }
-
-    // @@protoc_insertion_point(class_scope:google.protobuf.compiler.Version)
-    private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version DEFAULT_INSTANCE;
-    static {
-      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version();
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version getDefaultInstance() {
-      return DEFAULT_INSTANCE;
-    }
-
-    @java.lang.Deprecated public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Version>
-        PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Version>() {
-      public Version parsePartialFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-          return new Version(input, extensionRegistry);
-      }
-    };
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Version> parser() {
-      return PARSER;
-    }
-
-    @java.lang.Override
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Version> getParserForType() {
-      return PARSER;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version getDefaultInstanceForType() {
-      return DEFAULT_INSTANCE;
-    }
-
-  }
-
-  public interface CodeGeneratorRequestOrBuilder extends
-      // @@protoc_insertion_point(interface_extends:google.protobuf.compiler.CodeGeneratorRequest)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-    /**
-     * <pre>
-     * The .proto files that were explicitly listed on the command-line.  The
-     * code generator should generate code only for these files.  Each file's
-     * descriptor will be included in proto_file, below.
-     * </pre>
-     *
-     * <code>repeated string file_to_generate = 1;</code>
-     */
-    java.util.List<java.lang.String>
-        getFileToGenerateList();
-    /**
-     * <pre>
-     * The .proto files that were explicitly listed on the command-line.  The
-     * code generator should generate code only for these files.  Each file's
-     * descriptor will be included in proto_file, below.
-     * </pre>
-     *
-     * <code>repeated string file_to_generate = 1;</code>
-     */
-    int getFileToGenerateCount();
-    /**
-     * <pre>
-     * The .proto files that were explicitly listed on the command-line.  The
-     * code generator should generate code only for these files.  Each file's
-     * descriptor will be included in proto_file, below.
-     * </pre>
-     *
-     * <code>repeated string file_to_generate = 1;</code>
-     */
-    java.lang.String getFileToGenerate(int index);
-    /**
-     * <pre>
-     * The .proto files that were explicitly listed on the command-line.  The
-     * code generator should generate code only for these files.  Each file's
-     * descriptor will be included in proto_file, below.
-     * </pre>
-     *
-     * <code>repeated string file_to_generate = 1;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getFileToGenerateBytes(int index);
-
-    /**
-     * <pre>
-     * The generator parameter passed on the command-line.
-     * </pre>
-     *
-     * <code>optional string parameter = 2;</code>
-     */
-    boolean hasParameter();
-    /**
-     * <pre>
-     * The generator parameter passed on the command-line.
-     * </pre>
-     *
-     * <code>optional string parameter = 2;</code>
-     */
-    java.lang.String getParameter();
-    /**
-     * <pre>
-     * The generator parameter passed on the command-line.
-     * </pre>
-     *
-     * <code>optional string parameter = 2;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getParameterBytes();
-
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> 
-        getProtoFileList();
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto getProtoFile(int index);
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    int getProtoFileCount();
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder> 
-        getProtoFileOrBuilderList();
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder getProtoFileOrBuilder(
-        int index);
-
-    /**
-     * <pre>
-     * The version number of protocol compiler.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.compiler.Version compiler_version = 3;</code>
-     */
-    boolean hasCompilerVersion();
-    /**
-     * <pre>
-     * The version number of protocol compiler.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.compiler.Version compiler_version = 3;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version getCompilerVersion();
-    /**
-     * <pre>
-     * The version number of protocol compiler.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.compiler.Version compiler_version = 3;</code>
-     */
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.VersionOrBuilder getCompilerVersionOrBuilder();
-  }
-  /**
-   * <pre>
-   * An encoded CodeGeneratorRequest is written to the plugin's stdin.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.compiler.CodeGeneratorRequest}
-   */
-  public  static final class CodeGeneratorRequest extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-      // @@protoc_insertion_point(message_implements:google.protobuf.compiler.CodeGeneratorRequest)
-      CodeGeneratorRequestOrBuilder {
-    // Use CodeGeneratorRequest.newBuilder() to construct.
-    private CodeGeneratorRequest(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-      super(builder);
-    }
-    private CodeGeneratorRequest() {
-      fileToGenerate_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-      parameter_ = "";
-      protoFile_ = java.util.Collections.emptyList();
-    }
-
-    @java.lang.Override
-    public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-    getUnknownFields() {
-      return this.unknownFields;
-    }
-    private CodeGeneratorRequest(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      this();
-      int mutable_bitField0_ = 0;
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-                fileToGenerate_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList();
-                mutable_bitField0_ |= 0x00000001;
-              }
-              fileToGenerate_.add(bs);
-              break;
-            }
-            case 18: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = input.readBytes();
-              bitField0_ |= 0x00000001;
-              parameter_ = bs;
-              break;
-            }
-            case 26: {
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000002) == 0x00000002)) {
-                subBuilder = compilerVersion_.toBuilder();
-              }
-              compilerVersion_ = input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(compilerVersion_);
-                compilerVersion_ = subBuilder.buildPartial();
-              }
-              bitField0_ |= 0x00000002;
-              break;
-            }
-            case 122: {
-              if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-                protoFile_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto>();
-                mutable_bitField0_ |= 0x00000004;
-              }
-              protoFile_.add(
-                  input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.PARSER, extensionRegistry));
-              break;
-            }
-          }
-        }
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-            e).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-          fileToGenerate_ = fileToGenerate_.getUnmodifiableView();
-        }
-        if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-          protoFile_ = java.util.Collections.unmodifiableList(protoFile_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_CodeGeneratorRequest_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_CodeGeneratorRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest.Builder.class);
-    }
-
-    private int bitField0_;
-    public static final int FILE_TO_GENERATE_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList fileToGenerate_;
-    /**
-     * <pre>
-     * The .proto files that were explicitly listed on the command-line.  The
-     * code generator should generate code only for these files.  Each file's
-     * descriptor will be included in proto_file, below.
-     * </pre>
-     *
-     * <code>repeated string file_to_generate = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
-        getFileToGenerateList() {
-      return fileToGenerate_;
-    }
-    /**
-     * <pre>
-     * The .proto files that were explicitly listed on the command-line.  The
-     * code generator should generate code only for these files.  Each file's
-     * descriptor will be included in proto_file, below.
-     * </pre>
-     *
-     * <code>repeated string file_to_generate = 1;</code>
-     */
-    public int getFileToGenerateCount() {
-      return fileToGenerate_.size();
-    }
-    /**
-     * <pre>
-     * The .proto files that were explicitly listed on the command-line.  The
-     * code generator should generate code only for these files.  Each file's
-     * descriptor will be included in proto_file, below.
-     * </pre>
-     *
-     * <code>repeated string file_to_generate = 1;</code>
-     */
-    public java.lang.String getFileToGenerate(int index) {
-      return fileToGenerate_.get(index);
-    }
-    /**
-     * <pre>
-     * The .proto files that were explicitly listed on the command-line.  The
-     * code generator should generate code only for these files.  Each file's
-     * descriptor will be included in proto_file, below.
-     * </pre>
-     *
-     * <code>repeated string file_to_generate = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getFileToGenerateBytes(int index) {
-      return fileToGenerate_.getByteString(index);
-    }
-
-    public static final int PARAMETER_FIELD_NUMBER = 2;
-    private volatile java.lang.Object parameter_;
-    /**
-     * <pre>
-     * The generator parameter passed on the command-line.
-     * </pre>
-     *
-     * <code>optional string parameter = 2;</code>
-     */
-    public boolean hasParameter() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <pre>
-     * The generator parameter passed on the command-line.
-     * </pre>
-     *
-     * <code>optional string parameter = 2;</code>
-     */
-    public java.lang.String getParameter() {
-      java.lang.Object ref = parameter_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          parameter_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <pre>
-     * The generator parameter passed on the command-line.
-     * </pre>
-     *
-     * <code>optional string parameter = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getParameterBytes() {
-      java.lang.Object ref = parameter_;
-      if (ref instanceof java.lang.String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        parameter_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    public static final int PROTO_FILE_FIELD_NUMBER = 15;
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> protoFile_;
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> getProtoFileList() {
-      return protoFile_;
-    }
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder> 
-        getProtoFileOrBuilderList() {
-      return protoFile_;
-    }
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    public int getProtoFileCount() {
-      return protoFile_.size();
-    }
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto getProtoFile(int index) {
-      return protoFile_.get(index);
-    }
-    /**
-     * <pre>
-     * FileDescriptorProtos for all files in files_to_generate and everything
-     * they import.  The files will appear in topological order, so each file
-     * appears before any file that imports it.
-     * protoc guarantees that all proto_files will be written after
-     * the fields above, even though this is not technically guaranteed by the
-     * protobuf wire format.  This theoretically could allow a plugin to stream
-     * in the FileDescriptorProtos and handle them one by one rather than read
-     * the entire set into memory at once.  However, as of this writing, this
-     * is not similarly optimized on protoc's end -- it will store all fields in
-     * memory at once before sending them to the plugin.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.FileDescriptorProto proto_file = 15;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder getProtoFileOrBuilder(
-        int index) {
-      return protoFile_.get(index);
-    }
-
-    public static final int COMPILER_VERSION_FIELD_NUMBER = 3;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version compilerVersion_;
-    /**
-     * <pre>
-     * The version number of protocol compiler.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.compiler.Version compiler_version = 3;</code>
-     */
-    public boolean hasCompilerVersion() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <pre>
-     * The version number of protocol compiler.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.compiler.Version compiler_version = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version getCompilerVersion() {
-      return compilerVersion_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.getDefaultInstance() : compilerVersion_;
-    }
-    /**
-     * <pre>
-     * The version number of protocol compiler.
-     * </pre>
-     *
-     * <code>optional .google.protobuf.compiler.Version compiler_version = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.VersionOrBuilder getCompilerVersionOrBuilder() {
-      return compilerVersion_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.Version.getDefaultInstance() : compilerVersion_;
-    }
-
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized == 1) return true;
-      if (isInitialized == 0) return false;
-
-      for (int i = 0; i < getProtoFileCount(); i++) {
-        if (!getProtoFile(i).isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      for (int i = 0; i < fileToGenerate_.size(); i++) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, fileToGenerate_.getRaw(i));
-      }
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, parameter_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeMessage(3, getCompilerVersion());
-      }
-      for (int i = 0; i < protoFile_.size(); i++) {
-        output.writeMessage(15, protoFile_.get(i));
-      }
-      unknownFields.writeTo(output);
-    }
-
-    public int getSerializedSize() {
-      int size = memoizedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      {
-        int dataSize = 0;
-        for (int i = 0; i < fileToGenerate_.size(); i++) {
-          dataSize += computeStringSizeNoTag(fileToGenerate_.getRaw(i));
-        }
-        size += dataSize;
-        size += 1 * getFileToGenerateList().size();
-      }
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, parameter_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(3, getCompilerVersion());
-      }
-      for (int i = 0; i < protoFile_.size(); i++) {
-        size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(15, protoFile_.get(i));
-      }
-      size += unknownFields.getSerializedSize();
-      memoizedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest) obj;
-
-      boolean result = true;
-      result = result && getFileToGenerateList()
-          .equals(other.getFileToGenerateList());
-      result = result && (hasParameter() == other.hasParameter());
-      if (hasParameter()) {
-        result = result && getParameter()
-            .equals(other.getParameter());
-      }
-      result = result && getProtoFileList()
-          .equals(other.getProtoFileList());
-      result = result && (hasCompilerVersion() == other.hasCompilerVersion());
-      if (hasCompilerVersion()) {
-        result = result && getCompilerVersion()
-            .equals(other.getCompilerVersion());
-      }
-      result = result && unknownFields.equals(other.unknownFields);
-      return result;
-    }
-
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptor().hashCode();
-      if (getFileToGenerateCount() > 0) {
-        hash = (37 * hash) + FILE_TO_GENERATE_FIELD_NUMBER;
-        hash = (53 * hash) + getFileToGenerateList().hashCode();
-      }
-      if (hasParameter()) {
-        hash = (37 * hash) + PARAMETER_FIELD_NUMBER;
-        hash = (53 * hash) + getParameter().hashCode();
-      }
-      if (getProtoFileCount() > 0) {
-        hash = (37 * hash) + PROTO_FILE_FIELD_NUMBER;
-        hash = (53 * hash) + getProtoFileList().hashCode();
-      }
-      if (hasCompilerVersion()) {
-        hash = (37 * hash) + COMPILER_VERSION_FIELD_NUMBER;
-        hash = (53 * hash) + getCompilerVersion().hashCode();
-      }
-      hash = (29 * hash) + unknownFields.hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseFrom(byte[] data)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseFrom(
-        byte[] data,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input);
-    }
-    public static org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parseFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-          .parseWithIOException(PARSER, input, extensionRegistry);
-    }
-
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder() {
-      return DEFAULT_INSTANCE.toBuilder();
-    }
-    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest prototype) {
-      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() {
-      return this == DEFAULT_INSTANCE
-          ? new Builder() : new Builder().mergeFrom(this);
-    }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * <pre>
-     * An encoded CodeGeneratorRequest is written to the plugin's stdin.
-     * </pre>
-     *
-     * Protobuf type {@code google.protobuf.compiler.CodeGeneratorRequest}
-     */
-    public static final class Builder extends
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-        // @@protoc_insertion_point(builder_implements:google.protobuf.compiler.CodeGeneratorRequest)
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequestOrBuilder {
-      public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_CodeGeneratorRequest_descriptor;
-      }
-
-      protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_CodeGeneratorRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-                .alwaysUseFieldBuilders) {
-          getProtoFileFieldBuilder();
-          getCompilerVersionFieldBuilder();
-        }
-      }
-      public Builder clear() {
-        super.clear();
-        fileToGenerate_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        parameter_ = "";
-        bitField0_ = (bitField0_ & ~0x00000002);
-        if (protoFileBuilder_ == null) {
-          protoFile_ = java.util.Collections.emptyList();
-          bitField0_ = (bitField0_ & ~0x00000004);
-        } else {
-          protoFileBuilder_.clear();
-        }
-        if (compilerVersionBuilder_ == null) {
-          compilerVersion_ = null;
-        } else {
-          compilerVersionBuilder_.clear();
-        }
-        bitField0_ = (bitField0_ & ~0x00000008);
-        return this;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.internal_static_google_protobuf_compiler_CodeGeneratorRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest build() {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest buildPartial() {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((bitField0_ & 0x00000001) == 0x00000001)) {
-          fileToGenerate_ = fileToGenerate_.getUnmodifiableView();
-          bitField0_ = (bitField0_ & ~0x00000001);
-        }
-        result.fileToGenerate_ = fileToGenerate_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.parameter_ = parameter_;
-        if (protoFileBuilder_ == null) {
-          if (((bitField0_ & 0x00000004) == 0x00000004)) {
-            protoFile_ = java.util.Collections.unmodifiableList(protoFile_);
-            bitField0_ = (bitField0_ & ~0x00000004);
-          }
-          result.protoFile_ = protoFile_;
-        } else {
-          result.protoFile_ = protoFileBuilder_.build();
-        }
-        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        if (compilerVersionBuilder_ == null) {
-          result.compilerVersion_ = compilerVersion_;
-        } else {
-          result.compilerVersion_ = compilerVersionBuilder_.build();
-        }
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder clone() {
-        return (Builder) super.clone();
-      }
-      public Builder setField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.setField(field, value);
-      }
-      public Builder clearField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-        return (Builder) super.clearField(field);
-      }
-      public Builder clearOneof(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-        return (Builder) super.clearOneof(oneof);
-      }
-      public Builder setRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          int index, Object value) {
-        return (Builder) super.setRepeatedField(field, index, value);
-      }
-      public Builder addRepeatedField(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-          Object value) {
-        return (Builder) super.addRepeatedField(field, value);
-      }
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest other) {
-        if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest.getDefaultInstance()) return this;
-        if (!other.fileToGenerate_.isEmpty()) {
-          if (fileToGenerate_.isEmpty()) {
-            fileToGenerate_ = other.fileToGenerate_;
-            bitField0_ = (bitField0_ & ~0x00000001);
-          } else {
-            ensureFileToGenerateIsMutable();
-            fileToGenerate_.addAll(other.fileToGenerate_);
-          }
-          onChanged();
-        }
-        if (other.hasParameter()) {
-          bitField0_ |= 0x00000002;
-          parameter_ = other.parameter_;
-          onChanged();
-        }
-        if (protoFileBuilder_ == null) {
-          if (!other.protoFile_.isEmpty()) {
-            if (protoFile_.isEmpty()) {
-              protoFile_ = other.protoFile_;
-              bitField0_ = (bitField0_ & ~0x00000004);
-            } else {
-              ensureProtoFileIsMutable();
-              protoFile_.addAll(other.protoFile_);
-            }
-            onChanged();
-          }
-        } else {
-          if (!other.protoFile_.isEmpty()) {
-            if (protoFileBuilder_.isEmpty()) {
-              protoFileBuilder_.dispose();
-              protoFileBuilder_ = null;
-              protoFile_ = other.protoFile_;
-              bitField0_ = (bitField0_ & ~0x00000004);
-              protoFileBuilder_ = 
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                   getProtoFileFieldBuilder() : null;
-            } else {
-              protoFileBuilder_.addAllMessages(other.protoFile_);
-            }
-          }
-        }
-        if (other.hasCompilerVersion()) {
-          mergeCompilerVersion(other.getCompilerVersion());
-        }
-        this.mergeUnknownFields(other.unknownFields);
-        onChanged();
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        for (int i = 0; i < getProtoFileCount(); i++) {
-          if (!getProtoFile(i).isInitialized()) {
-            return false;
-          }
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.compiler.PluginProtos.CodeGeneratorRequest) e.getUnfinishedMessage();
-          throw e.unwrapIOException();
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList fileToGenerate_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-      private void ensureFileToGenerateIsMutable() {
-        if (!((bitField0_ & 0x00000001) == 0x00000001)) {
-          fileToGenerate_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList(fileToGenerate_);
-          bitField0_ |= 0x00000001;
-         }
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
-          getFileToGenerateList() {
-        return fileToGenerate_.getUnmodifiableView();
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public int getFileToGenerateCount() {
-        return fileToGenerate_.size();
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public java.lang.String getFileToGenerate(int index) {
-        return fileToGenerate_.get(index);
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-          getFileToGenerateBytes(int index) {
-        return fileToGenerate_.getByteString(index);
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public Builder setFileToGenerate(
-          int index, java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureFileToGenerateIsMutable();
-        fileToGenerate_.set(index, value);
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public Builder addFileToGenerate(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureFileToGenerateIsMutable();
-        fileToGenerate_.add(value);
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public Builder addAllFileToGenerate(
-          java.lang.Iterable<java.lang.String> values) {
-        ensureFileToGenerateIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, fileToGenerate_);
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public Builder clearFileToGenerate() {
-        fileToGenerate_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * The .proto files that were explicitly listed on the command-line.  The
-       * code generator should generate code only for these files.  Each file's
-       * descriptor will be included in proto_file, below.
-       * </pre>
-       *
-       * <code>repeated string file_to_generate = 1;</code>
-       */
-      public Builder addFileToGenerateBytes(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureFileToGenerateIsMutable();
-        fileToGenerate_.add(value);
-        onChanged();
-        return this;
-      }
-
-      private java.lang.Object parameter_ = "";
-      /**
-       * <pre>
-       * The generator parameter passed on the command-line.
-       * </pre>
-       *
-       * <code>optional string parameter = 2;</code>
-       */
-      public boolean hasParameter() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <pre>
-       * The generator parameter passed on the command-line.
-       * </pre>
-       *
-       * <code>optional string parameter = 2;</code>
-       */
-      public java.lang.String getParameter() {
-        java.lang.Object ref = parameter_;
-        if (!(ref instanceof java.lang.String)) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-              (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-          java.lang.String s = bs.toStringUtf8();
-          if (bs.isValidUtf8()) {
-            parameter_ = s;
-          }
-          return s;
-        } else {
-          return (java.lang.String) ref;
-        }
-      }
-      /**
-       * <pre>
-       * The generator parameter passed on the command-line.
-       * </pre>
-       *
-       * <code>optional string parameter = 2;</code>
-       */
-      public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-          getParameterBytes() {
-        java.lang.Object ref = parameter_;
-        if (ref instanceof String) {
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                  (java.lang.String) ref);
-          parameter_ = b;
-          return b;
-        } else {
-          return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        }
-      }
-      /**
-       * <pre>
-       * The generator parameter passed on the command-line.
-       * </pre>
-       *
-       * <code>optional string parameter = 2;</code>
-       */
-      public Builder setParameter(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000002;
-        parameter_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * The generator parameter passed on the command-line.
-       * </pre>
-       *
-       * <code>optional string parameter = 2;</code>
-       */
-      public Builder clearParameter() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        parameter_ = getDefaultInstance().getParameter();
-        onChanged();
-        return this;
-      }
-      /**
-       * <pre>
-       * The generator parameter passed on the command-line.
-       * </pre>
-       *
-       * <code>optional string parameter = 2;</code>
-       */
-      public Builder setParameterBytes(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000002;
-        parameter_ = value;
-        onChanged();
-        return this;
-      }
-
-      private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto> protoFile_ =
-        java.util.Collections.emptyList();
-      private void ensureProtoFileIsMutable() {
-        if (!((bitField0_ & 0x00000004) == 0x00000004)) {
-          protoFile_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto>(protoFile_);
-          bitField0_ |= 0x00000004;
-         }
-      }
-
-      private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProto.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.FileDescriptorProtoOrBuilder> protoFileBuilder_;
-
-      /**
-       * <pre>
-       * FileDescriptorProtos for all files in files_to_generate and everything
-       * they import.  The files will appear in topological order, so each file
-       * appears before any file that imports it.
-       * protoc guarantees that all proto_files will be written after
-       * the fields above, even though this is not technically guaranteed by the
-       * protobuf wire format.  This theoretically could allow a plugin to stream
-       * in the FileDescriptorProtos and handle them one by one rather than read
-       * the entire set into memory at once.  However, as of this writing, this
-       * is not similarly optimized on protoc's end -- it will store all fields in
-       * memory at once before sending them to the plugin.
-       * </pre>
-       *
-       * <code>repeated .google.protobuf.FileDescriptor

<TRUNCATED>

[31/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Empty.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Empty.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Empty.java
deleted file mode 100644
index 61f1574..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Empty.java
+++ /dev/null
@@ -1,386 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/empty.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * A generic empty message that you can re-use to avoid defining duplicated
- * empty messages in your APIs. A typical example is to use it as the request
- * or the response type of an API method. For instance:
- *     service Foo {
- *       rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty);
- *     }
- * The JSON representation for `Empty` is empty JSON object `{}`.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Empty}
- */
-public  final class Empty extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Empty)
-    EmptyOrBuilder {
-  // Use Empty.newBuilder() to construct.
-  private Empty(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Empty() {
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Empty(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.EmptyProto.internal_static_google_protobuf_Empty_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.EmptyProto.internal_static_google_protobuf_Empty_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty.Builder.class);
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty) obj;
-
-    boolean result = true;
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * A generic empty message that you can re-use to avoid defining duplicated
-   * empty messages in your APIs. A typical example is to use it as the request
-   * or the response type of an API method. For instance:
-   *     service Foo {
-   *       rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty);
-   *     }
-   * The JSON representation for `Empty` is empty JSON object `{}`.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Empty}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Empty)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.EmptyOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.EmptyProto.internal_static_google_protobuf_Empty_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.EmptyProto.internal_static_google_protobuf_Empty_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.EmptyProto.internal_static_google_protobuf_Empty_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty(this);
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty.getDefaultInstance()) return this;
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Empty)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Empty)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Empty>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Empty>() {
-    public Empty parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Empty(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Empty> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Empty> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Empty getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EmptyOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EmptyOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EmptyOrBuilder.java
deleted file mode 100644
index 15befc2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EmptyOrBuilder.java
+++ /dev/null
@@ -1,9 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/empty.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface EmptyOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Empty)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EmptyProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EmptyProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EmptyProto.java
deleted file mode 100644
index 5d4cb2a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EmptyProto.java
+++ /dev/null
@@ -1,58 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/empty.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class EmptyProto {
-  private EmptyProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Empty_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Empty_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\033google/protobuf/empty.proto\022\017google.pr" +
-      "otobuf\"\007\n\005EmptyBv\n\023com.google.protobufB\n" +
-      "EmptyProtoP\001Z\'github.com/golang/protobuf" +
-      "/ptypes/empty\370\001\001\242\002\003GPB\252\002\036Google.Protobuf" +
-      ".WellKnownTypesb\006proto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_google_protobuf_Empty_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_Empty_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Empty_descriptor,
-        new java.lang.String[] { });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Enum.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Enum.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Enum.java
deleted file mode 100644
index 3d2a206..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Enum.java
+++ /dev/null
@@ -1,1745 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Enum type definition.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Enum}
- */
-public  final class Enum extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Enum)
-    EnumOrBuilder {
-  // Use Enum.newBuilder() to construct.
-  private Enum(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Enum() {
-    name_ = "";
-    enumvalue_ = java.util.Collections.emptyList();
-    options_ = java.util.Collections.emptyList();
-    syntax_ = 0;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Enum(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            name_ = s;
-            break;
-          }
-          case 18: {
-            if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-              enumvalue_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue>();
-              mutable_bitField0_ |= 0x00000002;
-            }
-            enumvalue_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.parser(), extensionRegistry));
-            break;
-          }
-          case 26: {
-            if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-              options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>();
-              mutable_bitField0_ |= 0x00000004;
-            }
-            options_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.parser(), extensionRegistry));
-            break;
-          }
-          case 34: {
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder subBuilder = null;
-            if (sourceContext_ != null) {
-              subBuilder = sourceContext_.toBuilder();
-            }
-            sourceContext_ = input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.parser(), extensionRegistry);
-            if (subBuilder != null) {
-              subBuilder.mergeFrom(sourceContext_);
-              sourceContext_ = subBuilder.buildPartial();
-            }
-
-            break;
-          }
-          case 40: {
-            int rawValue = input.readEnum();
-
-            syntax_ = rawValue;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
-        enumvalue_ = java.util.Collections.unmodifiableList(enumvalue_);
-      }
-      if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-        options_ = java.util.Collections.unmodifiableList(options_);
-      }
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Enum_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Enum_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum.Builder.class);
-  }
-
-  private int bitField0_;
-  public static final int NAME_FIELD_NUMBER = 1;
-  private volatile java.lang.Object name_;
-  /**
-   * <pre>
-   * Enum type name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public java.lang.String getName() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      name_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * Enum type name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      name_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int ENUMVALUE_FIELD_NUMBER = 2;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue> enumvalue_;
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue> getEnumvalueList() {
-    return enumvalue_;
-  }
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder> 
-      getEnumvalueOrBuilderList() {
-    return enumvalue_;
-  }
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  public int getEnumvalueCount() {
-    return enumvalue_.size();
-  }
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue getEnumvalue(int index) {
-    return enumvalue_.get(index);
-  }
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder getEnumvalueOrBuilder(
-      int index) {
-    return enumvalue_.get(index);
-  }
-
-  public static final int OPTIONS_FIELD_NUMBER = 3;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_;
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public int getOptionsCount() {
-    return options_.size();
-  }
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-    return options_.get(index);
-  }
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index) {
-    return options_.get(index);
-  }
-
-  public static final int SOURCE_CONTEXT_FIELD_NUMBER = 4;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext sourceContext_;
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 4;</code>
-   */
-  public boolean hasSourceContext() {
-    return sourceContext_ != null;
-  }
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 4;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext() {
-    return sourceContext_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-  }
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 4;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder() {
-    return getSourceContext();
-  }
-
-  public static final int SYNTAX_FIELD_NUMBER = 5;
-  private int syntax_;
-  /**
-   * <pre>
-   * The source syntax.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 5;</code>
-   */
-  public int getSyntaxValue() {
-    return syntax_;
-  }
-  /**
-   * <pre>
-   * The source syntax.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 5;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax() {
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.valueOf(syntax_);
-    return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.UNRECOGNIZED : result;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
-    }
-    for (int i = 0; i < enumvalue_.size(); i++) {
-      output.writeMessage(2, enumvalue_.get(i));
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      output.writeMessage(3, options_.get(i));
-    }
-    if (sourceContext_ != null) {
-      output.writeMessage(4, getSourceContext());
-    }
-    if (syntax_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.SYNTAX_PROTO2.getNumber()) {
-      output.writeEnum(5, syntax_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
-    }
-    for (int i = 0; i < enumvalue_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(2, enumvalue_.get(i));
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(3, options_.get(i));
-    }
-    if (sourceContext_ != null) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(4, getSourceContext());
-    }
-    if (syntax_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.SYNTAX_PROTO2.getNumber()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeEnumSize(5, syntax_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum) obj;
-
-    boolean result = true;
-    result = result && getName()
-        .equals(other.getName());
-    result = result && getEnumvalueList()
-        .equals(other.getEnumvalueList());
-    result = result && getOptionsList()
-        .equals(other.getOptionsList());
-    result = result && (hasSourceContext() == other.hasSourceContext());
-    if (hasSourceContext()) {
-      result = result && getSourceContext()
-          .equals(other.getSourceContext());
-    }
-    result = result && syntax_ == other.syntax_;
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getName().hashCode();
-    if (getEnumvalueCount() > 0) {
-      hash = (37 * hash) + ENUMVALUE_FIELD_NUMBER;
-      hash = (53 * hash) + getEnumvalueList().hashCode();
-    }
-    if (getOptionsCount() > 0) {
-      hash = (37 * hash) + OPTIONS_FIELD_NUMBER;
-      hash = (53 * hash) + getOptionsList().hashCode();
-    }
-    if (hasSourceContext()) {
-      hash = (37 * hash) + SOURCE_CONTEXT_FIELD_NUMBER;
-      hash = (53 * hash) + getSourceContext().hashCode();
-    }
-    hash = (37 * hash) + SYNTAX_FIELD_NUMBER;
-    hash = (53 * hash) + syntax_;
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Enum type definition.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Enum}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Enum)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Enum_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Enum_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-        getEnumvalueFieldBuilder();
-        getOptionsFieldBuilder();
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      name_ = "";
-
-      if (enumvalueBuilder_ == null) {
-        enumvalue_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000002);
-      } else {
-        enumvalueBuilder_.clear();
-      }
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000004);
-      } else {
-        optionsBuilder_.clear();
-      }
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = null;
-      } else {
-        sourceContext_ = null;
-        sourceContextBuilder_ = null;
-      }
-      syntax_ = 0;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.internal_static_google_protobuf_Enum_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum(this);
-      int from_bitField0_ = bitField0_;
-      int to_bitField0_ = 0;
-      result.name_ = name_;
-      if (enumvalueBuilder_ == null) {
-        if (((bitField0_ & 0x00000002) == 0x00000002)) {
-          enumvalue_ = java.util.Collections.unmodifiableList(enumvalue_);
-          bitField0_ = (bitField0_ & ~0x00000002);
-        }
-        result.enumvalue_ = enumvalue_;
-      } else {
-        result.enumvalue_ = enumvalueBuilder_.build();
-      }
-      if (optionsBuilder_ == null) {
-        if (((bitField0_ & 0x00000004) == 0x00000004)) {
-          options_ = java.util.Collections.unmodifiableList(options_);
-          bitField0_ = (bitField0_ & ~0x00000004);
-        }
-        result.options_ = options_;
-      } else {
-        result.options_ = optionsBuilder_.build();
-      }
-      if (sourceContextBuilder_ == null) {
-        result.sourceContext_ = sourceContext_;
-      } else {
-        result.sourceContext_ = sourceContextBuilder_.build();
-      }
-      result.syntax_ = syntax_;
-      result.bitField0_ = to_bitField0_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum.getDefaultInstance()) return this;
-      if (!other.getName().isEmpty()) {
-        name_ = other.name_;
-        onChanged();
-      }
-      if (enumvalueBuilder_ == null) {
-        if (!other.enumvalue_.isEmpty()) {
-          if (enumvalue_.isEmpty()) {
-            enumvalue_ = other.enumvalue_;
-            bitField0_ = (bitField0_ & ~0x00000002);
-          } else {
-            ensureEnumvalueIsMutable();
-            enumvalue_.addAll(other.enumvalue_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.enumvalue_.isEmpty()) {
-          if (enumvalueBuilder_.isEmpty()) {
-            enumvalueBuilder_.dispose();
-            enumvalueBuilder_ = null;
-            enumvalue_ = other.enumvalue_;
-            bitField0_ = (bitField0_ & ~0x00000002);
-            enumvalueBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getEnumvalueFieldBuilder() : null;
-          } else {
-            enumvalueBuilder_.addAllMessages(other.enumvalue_);
-          }
-        }
-      }
-      if (optionsBuilder_ == null) {
-        if (!other.options_.isEmpty()) {
-          if (options_.isEmpty()) {
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000004);
-          } else {
-            ensureOptionsIsMutable();
-            options_.addAll(other.options_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.options_.isEmpty()) {
-          if (optionsBuilder_.isEmpty()) {
-            optionsBuilder_.dispose();
-            optionsBuilder_ = null;
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000004);
-            optionsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getOptionsFieldBuilder() : null;
-          } else {
-            optionsBuilder_.addAllMessages(other.options_);
-          }
-        }
-      }
-      if (other.hasSourceContext()) {
-        mergeSourceContext(other.getSourceContext());
-      }
-      if (other.syntax_ != 0) {
-        setSyntaxValue(other.getSyntaxValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private java.lang.Object name_ = "";
-    /**
-     * <pre>
-     * Enum type name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        name_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * Enum type name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * Enum type name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      name_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum type name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder clearName() {
-      
-      name_ = getDefaultInstance().getName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum type name.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      name_ = value;
-      onChanged();
-      return this;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue> enumvalue_ =
-      java.util.Collections.emptyList();
-    private void ensureEnumvalueIsMutable() {
-      if (!((bitField0_ & 0x00000002) == 0x00000002)) {
-        enumvalue_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue>(enumvalue_);
-        bitField0_ |= 0x00000002;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder> enumvalueBuilder_;
-
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue> getEnumvalueList() {
-      if (enumvalueBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(enumvalue_);
-      } else {
-        return enumvalueBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public int getEnumvalueCount() {
-      if (enumvalueBuilder_ == null) {
-        return enumvalue_.size();
-      } else {
-        return enumvalueBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue getEnumvalue(int index) {
-      if (enumvalueBuilder_ == null) {
-        return enumvalue_.get(index);
-      } else {
-        return enumvalueBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder setEnumvalue(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue value) {
-      if (enumvalueBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureEnumvalueIsMutable();
-        enumvalue_.set(index, value);
-        onChanged();
-      } else {
-        enumvalueBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder setEnumvalue(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder builderForValue) {
-      if (enumvalueBuilder_ == null) {
-        ensureEnumvalueIsMutable();
-        enumvalue_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        enumvalueBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder addEnumvalue(org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue value) {
-      if (enumvalueBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureEnumvalueIsMutable();
-        enumvalue_.add(value);
-        onChanged();
-      } else {
-        enumvalueBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder addEnumvalue(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue value) {
-      if (enumvalueBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureEnumvalueIsMutable();
-        enumvalue_.add(index, value);
-        onChanged();
-      } else {
-        enumvalueBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder addEnumvalue(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder builderForValue) {
-      if (enumvalueBuilder_ == null) {
-        ensureEnumvalueIsMutable();
-        enumvalue_.add(builderForValue.build());
-        onChanged();
-      } else {
-        enumvalueBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder addEnumvalue(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder builderForValue) {
-      if (enumvalueBuilder_ == null) {
-        ensureEnumvalueIsMutable();
-        enumvalue_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        enumvalueBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder addAllEnumvalue(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue> values) {
-      if (enumvalueBuilder_ == null) {
-        ensureEnumvalueIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, enumvalue_);
-        onChanged();
-      } else {
-        enumvalueBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder clearEnumvalue() {
-      if (enumvalueBuilder_ == null) {
-        enumvalue_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000002);
-        onChanged();
-      } else {
-        enumvalueBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public Builder removeEnumvalue(int index) {
-      if (enumvalueBuilder_ == null) {
-        ensureEnumvalueIsMutable();
-        enumvalue_.remove(index);
-        onChanged();
-      } else {
-        enumvalueBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder getEnumvalueBuilder(
-        int index) {
-      return getEnumvalueFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder getEnumvalueOrBuilder(
-        int index) {
-      if (enumvalueBuilder_ == null) {
-        return enumvalue_.get(index);  } else {
-        return enumvalueBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder> 
-         getEnumvalueOrBuilderList() {
-      if (enumvalueBuilder_ != null) {
-        return enumvalueBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(enumvalue_);
-      }
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder addEnumvalueBuilder() {
-      return getEnumvalueFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder addEnumvalueBuilder(
-        int index) {
-      return getEnumvalueFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Enum value definitions.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder> 
-         getEnumvalueBuilderList() {
-      return getEnumvalueFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder> 
-        getEnumvalueFieldBuilder() {
-      if (enumvalueBuilder_ == null) {
-        enumvalueBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder>(
-                enumvalue_,
-                ((bitField0_ & 0x00000002) == 0x00000002),
-                getParentForChildren(),
-                isClean());
-        enumvalue_ = null;
-      }
-      return enumvalueBuilder_;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_ =
-      java.util.Collections.emptyList();
-    private void ensureOptionsIsMutable() {
-      if (!((bitField0_ & 0x00000004) == 0x00000004)) {
-        options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>(options_);
-        bitField0_ |= 0x00000004;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> optionsBuilder_;
-
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-      if (optionsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(options_);
-      } else {
-        return optionsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public int getOptionsCount() {
-      if (optionsBuilder_ == null) {
-        return options_.size();
-      } else {
-        return optionsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);
-      } else {
-        return optionsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.set(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder addAllOptions(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> values) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, options_);
-        onChanged();
-      } else {
-        optionsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder clearOptions() {
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000004);
-        onChanged();
-      } else {
-        optionsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public Builder removeOptions(int index) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.remove(index);
-        onChanged();
-      } else {
-        optionsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder getOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-        int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);  } else {
-        return optionsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-         getOptionsOrBuilderList() {
-      if (optionsBuilder_ != null) {
-        return optionsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(options_);
-      }
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder() {
-      return getOptionsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Protocol buffer options.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 3;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder> 
-         getOptionsBuilderList() {
-      return getOptionsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-        getOptionsFieldBuilder() {
-      if (optionsBuilder_ == null) {
-        optionsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder>(
-                options_,
-                ((bitField0_ & 0x00000004) == 0x00000004),
-                getParentForChildren(),
-                isClean());
-        options_ = null;
-      }
-      return optionsBuilder_;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext sourceContext_ = null;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder> sourceContextBuilder_;
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    public boolean hasSourceContext() {
-      return sourceContextBuilder_ != null || sourceContext_ != null;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext() {
-      if (sourceContextBuilder_ == null) {
-        return sourceContext_ == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-      } else {
-        return sourceContextBuilder_.getMessage();
-      }
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    public Builder setSourceContext(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext value) {
-      if (sourceContextBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        sourceContext_ = value;
-        onChanged();
-      } else {
-        sourceContextBuilder_.setMessage(value);
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    public Builder setSourceContext(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder builderForValue) {
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = builderForValue.build();
-        onChanged();
-      } else {
-        sourceContextBuilder_.setMessage(builderForValue.build());
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    public Builder mergeSourceContext(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext value) {
-      if (sourceContextBuilder_ == null) {
-        if (sourceContext_ != null) {
-          sourceContext_ =
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.newBuilder(sourceContext_).mergeFrom(value).buildPartial();
-        } else {
-          sourceContext_ = value;
-        }
-        onChanged();
-      } else {
-        sourceContextBuilder_.mergeFrom(value);
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    public Builder clearSourceContext() {
-      if (sourceContextBuilder_ == null) {
-        sourceContext_ = null;
-        onChanged();
-      } else {
-        sourceContext_ = null;
-        sourceContextBuilder_ = null;
-      }
-
-      return this;
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder getSourceContextBuilder() {
-      
-      onChanged();
-      return getSourceContextFieldBuilder().getBuilder();
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder() {
-      if (sourceContextBuilder_ != null) {
-        return sourceContextBuilder_.getMessageOrBuilder();
-      } else {
-        return sourceContext_ == null ?
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance() : sourceContext_;
-      }
-    }
-    /**
-     * <pre>
-     * The source context.
-     * </pre>
-     *
-     * <code>.google.protobuf.SourceContext source_context = 4;</code>
-     */
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder> 
-        getSourceContextFieldBuilder() {
-      if (sourceContextBuilder_ == null) {
-        sourceContextBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SingleFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder>(
-                getSourceContext(),
-                getParentForChildren(),
-                isClean());
-        sourceContext_ = null;
-      }
-      return sourceContextBuilder_;
-    }
-
-    private int syntax_ = 0;
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 5;</code>
-     */
-    public int getSyntaxValue() {
-      return syntax_;
-    }
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 5;</code>
-     */
-    public Builder setSyntaxValue(int value) {
-      syntax_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 5;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.valueOf(syntax_);
-      return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.UNRECOGNIZED : result;
-    }
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 5;</code>
-     */
-    public Builder setSyntax(org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax value) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-      
-      syntax_ = value.getNumber();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The source syntax.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 5;</code>
-     */
-    public Builder clearSyntax() {
-      
-      syntax_ = 0;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Enum)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Enum)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Enum>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Enum>() {
-    public Enum parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Enum(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Enum> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Enum> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Enum getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumOrBuilder.java
deleted file mode 100644
index 6e2d7d4..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/EnumOrBuilder.java
+++ /dev/null
@@ -1,157 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface EnumOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Enum)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * Enum type name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  java.lang.String getName();
-  /**
-   * <pre>
-   * Enum type name.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes();
-
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue> 
-      getEnumvalueList();
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValue getEnumvalue(int index);
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  int getEnumvalueCount();
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder> 
-      getEnumvalueOrBuilderList();
-  /**
-   * <pre>
-   * Enum value definitions.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.EnumValue enumvalue = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.EnumValueOrBuilder getEnumvalueOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> 
-      getOptionsList();
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index);
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  int getOptionsCount();
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList();
-  /**
-   * <pre>
-   * Protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 3;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 4;</code>
-   */
-  boolean hasSourceContext();
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 4;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext();
-  /**
-   * <pre>
-   * The source context.
-   * </pre>
-   *
-   * <code>.google.protobuf.SourceContext source_context = 4;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder();
-
-  /**
-   * <pre>
-   * The source syntax.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 5;</code>
-   */
-  int getSyntaxValue();
-  /**
-   * <pre>
-   * The source syntax.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 5;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax();
-}


[19/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Method.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Method.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Method.java
deleted file mode 100644
index 32e8b35..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Method.java
+++ /dev/null
@@ -1,1541 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Method represents a method of an api.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Method}
- */
-public  final class Method extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Method)
-    MethodOrBuilder {
-  // Use Method.newBuilder() to construct.
-  private Method(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Method() {
-    name_ = "";
-    requestTypeUrl_ = "";
-    requestStreaming_ = false;
-    responseTypeUrl_ = "";
-    responseStreaming_ = false;
-    options_ = java.util.Collections.emptyList();
-    syntax_ = 0;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Method(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            name_ = s;
-            break;
-          }
-          case 18: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            requestTypeUrl_ = s;
-            break;
-          }
-          case 24: {
-
-            requestStreaming_ = input.readBool();
-            break;
-          }
-          case 34: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            responseTypeUrl_ = s;
-            break;
-          }
-          case 40: {
-
-            responseStreaming_ = input.readBool();
-            break;
-          }
-          case 50: {
-            if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
-              options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>();
-              mutable_bitField0_ |= 0x00000020;
-            }
-            options_.add(
-                input.readMessage(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.parser(), extensionRegistry));
-            break;
-          }
-          case 56: {
-            int rawValue = input.readEnum();
-
-            syntax_ = rawValue;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
-        options_ = java.util.Collections.unmodifiableList(options_);
-      }
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Method_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Method_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder.class);
-  }
-
-  private int bitField0_;
-  public static final int NAME_FIELD_NUMBER = 1;
-  private volatile java.lang.Object name_;
-  /**
-   * <pre>
-   * The simple name of this method.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public java.lang.String getName() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      name_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The simple name of this method.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      name_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int REQUEST_TYPE_URL_FIELD_NUMBER = 2;
-  private volatile java.lang.Object requestTypeUrl_;
-  /**
-   * <pre>
-   * A URL of the input message type.
-   * </pre>
-   *
-   * <code>string request_type_url = 2;</code>
-   */
-  public java.lang.String getRequestTypeUrl() {
-    java.lang.Object ref = requestTypeUrl_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      requestTypeUrl_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * A URL of the input message type.
-   * </pre>
-   *
-   * <code>string request_type_url = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getRequestTypeUrlBytes() {
-    java.lang.Object ref = requestTypeUrl_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      requestTypeUrl_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int REQUEST_STREAMING_FIELD_NUMBER = 3;
-  private boolean requestStreaming_;
-  /**
-   * <pre>
-   * If true, the request is streamed.
-   * </pre>
-   *
-   * <code>bool request_streaming = 3;</code>
-   */
-  public boolean getRequestStreaming() {
-    return requestStreaming_;
-  }
-
-  public static final int RESPONSE_TYPE_URL_FIELD_NUMBER = 4;
-  private volatile java.lang.Object responseTypeUrl_;
-  /**
-   * <pre>
-   * The URL of the output message type.
-   * </pre>
-   *
-   * <code>string response_type_url = 4;</code>
-   */
-  public java.lang.String getResponseTypeUrl() {
-    java.lang.Object ref = responseTypeUrl_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      responseTypeUrl_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The URL of the output message type.
-   * </pre>
-   *
-   * <code>string response_type_url = 4;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getResponseTypeUrlBytes() {
-    java.lang.Object ref = responseTypeUrl_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      responseTypeUrl_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int RESPONSE_STREAMING_FIELD_NUMBER = 5;
-  private boolean responseStreaming_;
-  /**
-   * <pre>
-   * If true, the response is streamed.
-   * </pre>
-   *
-   * <code>bool response_streaming = 5;</code>
-   */
-  public boolean getResponseStreaming() {
-    return responseStreaming_;
-  }
-
-  public static final int OPTIONS_FIELD_NUMBER = 6;
-  private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_;
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList() {
-    return options_;
-  }
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  public int getOptionsCount() {
-    return options_.size();
-  }
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-    return options_.get(index);
-  }
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index) {
-    return options_.get(index);
-  }
-
-  public static final int SYNTAX_FIELD_NUMBER = 7;
-  private int syntax_;
-  /**
-   * <pre>
-   * The source syntax of this method.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 7;</code>
-   */
-  public int getSyntaxValue() {
-    return syntax_;
-  }
-  /**
-   * <pre>
-   * The source syntax of this method.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 7;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax() {
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.valueOf(syntax_);
-    return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.UNRECOGNIZED : result;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
-    }
-    if (!getRequestTypeUrlBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, requestTypeUrl_);
-    }
-    if (requestStreaming_ != false) {
-      output.writeBool(3, requestStreaming_);
-    }
-    if (!getResponseTypeUrlBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, responseTypeUrl_);
-    }
-    if (responseStreaming_ != false) {
-      output.writeBool(5, responseStreaming_);
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      output.writeMessage(6, options_.get(i));
-    }
-    if (syntax_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.SYNTAX_PROTO2.getNumber()) {
-      output.writeEnum(7, syntax_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
-    }
-    if (!getRequestTypeUrlBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, requestTypeUrl_);
-    }
-    if (requestStreaming_ != false) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeBoolSize(3, requestStreaming_);
-    }
-    if (!getResponseTypeUrlBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, responseTypeUrl_);
-    }
-    if (responseStreaming_ != false) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeBoolSize(5, responseStreaming_);
-    }
-    for (int i = 0; i < options_.size(); i++) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeMessageSize(6, options_.get(i));
-    }
-    if (syntax_ != org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.SYNTAX_PROTO2.getNumber()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeEnumSize(7, syntax_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Method)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Method other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Method) obj;
-
-    boolean result = true;
-    result = result && getName()
-        .equals(other.getName());
-    result = result && getRequestTypeUrl()
-        .equals(other.getRequestTypeUrl());
-    result = result && (getRequestStreaming()
-        == other.getRequestStreaming());
-    result = result && getResponseTypeUrl()
-        .equals(other.getResponseTypeUrl());
-    result = result && (getResponseStreaming()
-        == other.getResponseStreaming());
-    result = result && getOptionsList()
-        .equals(other.getOptionsList());
-    result = result && syntax_ == other.syntax_;
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getName().hashCode();
-    hash = (37 * hash) + REQUEST_TYPE_URL_FIELD_NUMBER;
-    hash = (53 * hash) + getRequestTypeUrl().hashCode();
-    hash = (37 * hash) + REQUEST_STREAMING_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-        getRequestStreaming());
-    hash = (37 * hash) + RESPONSE_TYPE_URL_FIELD_NUMBER;
-    hash = (53 * hash) + getResponseTypeUrl().hashCode();
-    hash = (37 * hash) + RESPONSE_STREAMING_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean(
-        getResponseStreaming());
-    if (getOptionsCount() > 0) {
-      hash = (37 * hash) + OPTIONS_FIELD_NUMBER;
-      hash = (53 * hash) + getOptionsList().hashCode();
-    }
-    hash = (37 * hash) + SYNTAX_FIELD_NUMBER;
-    hash = (53 * hash) + syntax_;
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Method prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Method represents a method of an api.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Method}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Method)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Method_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Method_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-        getOptionsFieldBuilder();
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      name_ = "";
-
-      requestTypeUrl_ = "";
-
-      requestStreaming_ = false;
-
-      responseTypeUrl_ = "";
-
-      responseStreaming_ = false;
-
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000020);
-      } else {
-        optionsBuilder_.clear();
-      }
-      syntax_ = 0;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Method_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Method result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Method result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Method(this);
-      int from_bitField0_ = bitField0_;
-      int to_bitField0_ = 0;
-      result.name_ = name_;
-      result.requestTypeUrl_ = requestTypeUrl_;
-      result.requestStreaming_ = requestStreaming_;
-      result.responseTypeUrl_ = responseTypeUrl_;
-      result.responseStreaming_ = responseStreaming_;
-      if (optionsBuilder_ == null) {
-        if (((bitField0_ & 0x00000020) == 0x00000020)) {
-          options_ = java.util.Collections.unmodifiableList(options_);
-          bitField0_ = (bitField0_ & ~0x00000020);
-        }
-        result.options_ = options_;
-      } else {
-        result.options_ = optionsBuilder_.build();
-      }
-      result.syntax_ = syntax_;
-      result.bitField0_ = to_bitField0_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Method) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Method)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Method other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Method.getDefaultInstance()) return this;
-      if (!other.getName().isEmpty()) {
-        name_ = other.name_;
-        onChanged();
-      }
-      if (!other.getRequestTypeUrl().isEmpty()) {
-        requestTypeUrl_ = other.requestTypeUrl_;
-        onChanged();
-      }
-      if (other.getRequestStreaming() != false) {
-        setRequestStreaming(other.getRequestStreaming());
-      }
-      if (!other.getResponseTypeUrl().isEmpty()) {
-        responseTypeUrl_ = other.responseTypeUrl_;
-        onChanged();
-      }
-      if (other.getResponseStreaming() != false) {
-        setResponseStreaming(other.getResponseStreaming());
-      }
-      if (optionsBuilder_ == null) {
-        if (!other.options_.isEmpty()) {
-          if (options_.isEmpty()) {
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000020);
-          } else {
-            ensureOptionsIsMutable();
-            options_.addAll(other.options_);
-          }
-          onChanged();
-        }
-      } else {
-        if (!other.options_.isEmpty()) {
-          if (optionsBuilder_.isEmpty()) {
-            optionsBuilder_.dispose();
-            optionsBuilder_ = null;
-            options_ = other.options_;
-            bitField0_ = (bitField0_ & ~0x00000020);
-            optionsBuilder_ = 
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
-                 getOptionsFieldBuilder() : null;
-          } else {
-            optionsBuilder_.addAllMessages(other.options_);
-          }
-        }
-      }
-      if (other.syntax_ != 0) {
-        setSyntaxValue(other.getSyntaxValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Method parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Method) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private java.lang.Object name_ = "";
-    /**
-     * <pre>
-     * The simple name of this method.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        name_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The simple name of this method.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The simple name of this method.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      name_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The simple name of this method.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder clearName() {
-      
-      name_ = getDefaultInstance().getName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The simple name of this method.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      name_ = value;
-      onChanged();
-      return this;
-    }
-
-    private java.lang.Object requestTypeUrl_ = "";
-    /**
-     * <pre>
-     * A URL of the input message type.
-     * </pre>
-     *
-     * <code>string request_type_url = 2;</code>
-     */
-    public java.lang.String getRequestTypeUrl() {
-      java.lang.Object ref = requestTypeUrl_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        requestTypeUrl_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * A URL of the input message type.
-     * </pre>
-     *
-     * <code>string request_type_url = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getRequestTypeUrlBytes() {
-      java.lang.Object ref = requestTypeUrl_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        requestTypeUrl_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * A URL of the input message type.
-     * </pre>
-     *
-     * <code>string request_type_url = 2;</code>
-     */
-    public Builder setRequestTypeUrl(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      requestTypeUrl_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * A URL of the input message type.
-     * </pre>
-     *
-     * <code>string request_type_url = 2;</code>
-     */
-    public Builder clearRequestTypeUrl() {
-      
-      requestTypeUrl_ = getDefaultInstance().getRequestTypeUrl();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * A URL of the input message type.
-     * </pre>
-     *
-     * <code>string request_type_url = 2;</code>
-     */
-    public Builder setRequestTypeUrlBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      requestTypeUrl_ = value;
-      onChanged();
-      return this;
-    }
-
-    private boolean requestStreaming_ ;
-    /**
-     * <pre>
-     * If true, the request is streamed.
-     * </pre>
-     *
-     * <code>bool request_streaming = 3;</code>
-     */
-    public boolean getRequestStreaming() {
-      return requestStreaming_;
-    }
-    /**
-     * <pre>
-     * If true, the request is streamed.
-     * </pre>
-     *
-     * <code>bool request_streaming = 3;</code>
-     */
-    public Builder setRequestStreaming(boolean value) {
-      
-      requestStreaming_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * If true, the request is streamed.
-     * </pre>
-     *
-     * <code>bool request_streaming = 3;</code>
-     */
-    public Builder clearRequestStreaming() {
-      
-      requestStreaming_ = false;
-      onChanged();
-      return this;
-    }
-
-    private java.lang.Object responseTypeUrl_ = "";
-    /**
-     * <pre>
-     * The URL of the output message type.
-     * </pre>
-     *
-     * <code>string response_type_url = 4;</code>
-     */
-    public java.lang.String getResponseTypeUrl() {
-      java.lang.Object ref = responseTypeUrl_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        responseTypeUrl_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The URL of the output message type.
-     * </pre>
-     *
-     * <code>string response_type_url = 4;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getResponseTypeUrlBytes() {
-      java.lang.Object ref = responseTypeUrl_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        responseTypeUrl_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The URL of the output message type.
-     * </pre>
-     *
-     * <code>string response_type_url = 4;</code>
-     */
-    public Builder setResponseTypeUrl(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      responseTypeUrl_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The URL of the output message type.
-     * </pre>
-     *
-     * <code>string response_type_url = 4;</code>
-     */
-    public Builder clearResponseTypeUrl() {
-      
-      responseTypeUrl_ = getDefaultInstance().getResponseTypeUrl();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The URL of the output message type.
-     * </pre>
-     *
-     * <code>string response_type_url = 4;</code>
-     */
-    public Builder setResponseTypeUrlBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      responseTypeUrl_ = value;
-      onChanged();
-      return this;
-    }
-
-    private boolean responseStreaming_ ;
-    /**
-     * <pre>
-     * If true, the response is streamed.
-     * </pre>
-     *
-     * <code>bool response_streaming = 5;</code>
-     */
-    public boolean getResponseStreaming() {
-      return responseStreaming_;
-    }
-    /**
-     * <pre>
-     * If true, the response is streamed.
-     * </pre>
-     *
-     * <code>bool response_streaming = 5;</code>
-     */
-    public Builder setResponseStreaming(boolean value) {
-      
-      responseStreaming_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * If true, the response is streamed.
-     * </pre>
-     *
-     * <code>bool response_streaming = 5;</code>
-     */
-    public Builder clearResponseStreaming() {
-      
-      responseStreaming_ = false;
-      onChanged();
-      return this;
-    }
-
-    private java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> options_ =
-      java.util.Collections.emptyList();
-    private void ensureOptionsIsMutable() {
-      if (!((bitField0_ & 0x00000020) == 0x00000020)) {
-        options_ = new java.util.ArrayList<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option>(options_);
-        bitField0_ |= 0x00000020;
-       }
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> optionsBuilder_;
-
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> getOptionsList() {
-      if (optionsBuilder_ == null) {
-        return java.util.Collections.unmodifiableList(options_);
-      } else {
-        return optionsBuilder_.getMessageList();
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public int getOptionsCount() {
-      if (optionsBuilder_ == null) {
-        return options_.size();
-      } else {
-        return optionsBuilder_.getCount();
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);
-      } else {
-        return optionsBuilder_.getMessage(index);
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.set(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder setOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.set(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.setMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder addOptions(org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option value) {
-      if (optionsBuilder_ == null) {
-        if (value == null) {
-          throw new NullPointerException();
-        }
-        ensureOptionsIsMutable();
-        options_.add(index, value);
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, value);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder addOptions(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder addOptions(
-        int index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder builderForValue) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.add(index, builderForValue.build());
-        onChanged();
-      } else {
-        optionsBuilder_.addMessage(index, builderForValue.build());
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder addAllOptions(
-        java.lang.Iterable<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> values) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-            values, options_);
-        onChanged();
-      } else {
-        optionsBuilder_.addAllMessages(values);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder clearOptions() {
-      if (optionsBuilder_ == null) {
-        options_ = java.util.Collections.emptyList();
-        bitField0_ = (bitField0_ & ~0x00000020);
-        onChanged();
-      } else {
-        optionsBuilder_.clear();
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public Builder removeOptions(int index) {
-      if (optionsBuilder_ == null) {
-        ensureOptionsIsMutable();
-        options_.remove(index);
-        onChanged();
-      } else {
-        optionsBuilder_.remove(index);
-      }
-      return this;
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder getOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().getBuilder(index);
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-        int index) {
-      if (optionsBuilder_ == null) {
-        return options_.get(index);  } else {
-        return optionsBuilder_.getMessageOrBuilder(index);
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-         getOptionsOrBuilderList() {
-      if (optionsBuilder_ != null) {
-        return optionsBuilder_.getMessageOrBuilderList();
-      } else {
-        return java.util.Collections.unmodifiableList(options_);
-      }
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder() {
-      return getOptionsFieldBuilder().addBuilder(
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder addOptionsBuilder(
-        int index) {
-      return getOptionsFieldBuilder().addBuilder(
-          index, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.getDefaultInstance());
-    }
-    /**
-     * <pre>
-     * Any metadata attached to the method.
-     * </pre>
-     *
-     * <code>repeated .google.protobuf.Option options = 6;</code>
-     */
-    public java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder> 
-         getOptionsBuilderList() {
-      return getOptionsFieldBuilder().getBuilderList();
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-        getOptionsFieldBuilder() {
-      if (optionsBuilder_ == null) {
-        optionsBuilder_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.RepeatedFieldBuilderV3<
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Option, org.apache.hadoop.hbase.shaded.com.google.protobuf.Option.Builder, org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder>(
-                options_,
-                ((bitField0_ & 0x00000020) == 0x00000020),
-                getParentForChildren(),
-                isClean());
-        options_ = null;
-      }
-      return optionsBuilder_;
-    }
-
-    private int syntax_ = 0;
-    /**
-     * <pre>
-     * The source syntax of this method.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public int getSyntaxValue() {
-      return syntax_;
-    }
-    /**
-     * <pre>
-     * The source syntax of this method.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public Builder setSyntaxValue(int value) {
-      syntax_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The source syntax of this method.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax result = org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.valueOf(syntax_);
-      return result == null ? org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax.UNRECOGNIZED : result;
-    }
-    /**
-     * <pre>
-     * The source syntax of this method.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public Builder setSyntax(org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax value) {
-      if (value == null) {
-        throw new NullPointerException();
-      }
-      
-      syntax_ = value.getNumber();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The source syntax of this method.
-     * </pre>
-     *
-     * <code>.google.protobuf.Syntax syntax = 7;</code>
-     */
-    public Builder clearSyntax() {
-      
-      syntax_ = 0;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Method)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Method)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Method DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Method();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Method getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Method>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Method>() {
-    public Method parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Method(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Method> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Method> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Method getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MethodOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MethodOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MethodOrBuilder.java
deleted file mode 100644
index 12819df..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MethodOrBuilder.java
+++ /dev/null
@@ -1,142 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface MethodOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Method)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The simple name of this method.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  java.lang.String getName();
-  /**
-   * <pre>
-   * The simple name of this method.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes();
-
-  /**
-   * <pre>
-   * A URL of the input message type.
-   * </pre>
-   *
-   * <code>string request_type_url = 2;</code>
-   */
-  java.lang.String getRequestTypeUrl();
-  /**
-   * <pre>
-   * A URL of the input message type.
-   * </pre>
-   *
-   * <code>string request_type_url = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getRequestTypeUrlBytes();
-
-  /**
-   * <pre>
-   * If true, the request is streamed.
-   * </pre>
-   *
-   * <code>bool request_streaming = 3;</code>
-   */
-  boolean getRequestStreaming();
-
-  /**
-   * <pre>
-   * The URL of the output message type.
-   * </pre>
-   *
-   * <code>string response_type_url = 4;</code>
-   */
-  java.lang.String getResponseTypeUrl();
-  /**
-   * <pre>
-   * The URL of the output message type.
-   * </pre>
-   *
-   * <code>string response_type_url = 4;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getResponseTypeUrlBytes();
-
-  /**
-   * <pre>
-   * If true, the response is streamed.
-   * </pre>
-   *
-   * <code>bool response_streaming = 5;</code>
-   */
-  boolean getResponseStreaming();
-
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> 
-      getOptionsList();
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index);
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  int getOptionsCount();
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList();
-  /**
-   * <pre>
-   * Any metadata attached to the method.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 6;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * The source syntax of this method.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 7;</code>
-   */
-  int getSyntaxValue();
-  /**
-   * <pre>
-   * The source syntax of this method.
-   * </pre>
-   *
-   * <code>.google.protobuf.Syntax syntax = 7;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Mixin.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Mixin.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Mixin.java
deleted file mode 100644
index 5dd6be6..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Mixin.java
+++ /dev/null
@@ -1,812 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Declares an API to be included in this API. The including API must
- * redeclare all the methods from the included API, but documentation
- * and options are inherited as follows:
- * - If after comment and whitespace stripping, the documentation
- *   string of the redeclared method is empty, it will be inherited
- *   from the original method.
- * - Each annotation belonging to the service config (http,
- *   visibility) which is not set in the redeclared method will be
- *   inherited.
- * - If an http annotation is inherited, the path pattern will be
- *   modified as follows. Any version prefix will be replaced by the
- *   version of the including API plus the [root][] path if specified.
- * Example of a simple mixin:
- *     package google.acl.v1;
- *     service AccessControl {
- *       // Get the underlying ACL object.
- *       rpc GetAcl(GetAclRequest) returns (Acl) {
- *         option (google.api.http).get = "/v1/{resource=**}:getAcl";
- *       }
- *     }
- *     package google.storage.v2;
- *     service Storage {
- *       rpc GetAcl(GetAclRequest) returns (Acl);
- *       // Get a data record.
- *       rpc GetData(GetDataRequest) returns (Data) {
- *         option (google.api.http).get = "/v2/{resource=**}";
- *       }
- *     }
- * Example of a mixin configuration:
- *     apis:
- *     - name: google.storage.v2.Storage
- *       mixins:
- *       - name: google.acl.v1.AccessControl
- * The mixin construct implies that all methods in `AccessControl` are
- * also declared with same name and request/response types in
- * `Storage`. A documentation generator or annotation processor will
- * see the effective `Storage.GetAcl` method after inherting
- * documentation and annotations as follows:
- *     service Storage {
- *       // Get the underlying ACL object.
- *       rpc GetAcl(GetAclRequest) returns (Acl) {
- *         option (google.api.http).get = "/v2/{resource=**}:getAcl";
- *       }
- *       ...
- *     }
- * Note how the version in the path pattern changed from `v1` to `v2`.
- * If the `root` field in the mixin is specified, it should be a
- * relative path under which inherited HTTP paths are placed. Example:
- *     apis:
- *     - name: google.storage.v2.Storage
- *       mixins:
- *       - name: google.acl.v1.AccessControl
- *         root: acls
- * This implies the following inherited HTTP annotation:
- *     service Storage {
- *       // Get the underlying ACL object.
- *       rpc GetAcl(GetAclRequest) returns (Acl) {
- *         option (google.api.http).get = "/v2/acls/{resource=**}:getAcl";
- *       }
- *       ...
- *     }
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Mixin}
- */
-public  final class Mixin extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Mixin)
-    MixinOrBuilder {
-  // Use Mixin.newBuilder() to construct.
-  private Mixin(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Mixin() {
-    name_ = "";
-    root_ = "";
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Mixin(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            name_ = s;
-            break;
-          }
-          case 18: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            root_ = s;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Mixin_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Mixin_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder.class);
-  }
-
-  public static final int NAME_FIELD_NUMBER = 1;
-  private volatile java.lang.Object name_;
-  /**
-   * <pre>
-   * The fully qualified name of the API which is included.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public java.lang.String getName() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      name_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The fully qualified name of the API which is included.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes() {
-    java.lang.Object ref = name_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      name_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int ROOT_FIELD_NUMBER = 2;
-  private volatile java.lang.Object root_;
-  /**
-   * <pre>
-   * If non-empty specifies a path under which inherited HTTP paths
-   * are rooted.
-   * </pre>
-   *
-   * <code>string root = 2;</code>
-   */
-  public java.lang.String getRoot() {
-    java.lang.Object ref = root_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      root_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * If non-empty specifies a path under which inherited HTTP paths
-   * are rooted.
-   * </pre>
-   *
-   * <code>string root = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getRootBytes() {
-    java.lang.Object ref = root_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      root_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, name_);
-    }
-    if (!getRootBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, root_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, name_);
-    }
-    if (!getRootBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, root_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin) obj;
-
-    boolean result = true;
-    result = result && getName()
-        .equals(other.getName());
-    result = result && getRoot()
-        .equals(other.getRoot());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getName().hashCode();
-    hash = (37 * hash) + ROOT_FIELD_NUMBER;
-    hash = (53 * hash) + getRoot().hashCode();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Declares an API to be included in this API. The including API must
-   * redeclare all the methods from the included API, but documentation
-   * and options are inherited as follows:
-   * - If after comment and whitespace stripping, the documentation
-   *   string of the redeclared method is empty, it will be inherited
-   *   from the original method.
-   * - Each annotation belonging to the service config (http,
-   *   visibility) which is not set in the redeclared method will be
-   *   inherited.
-   * - If an http annotation is inherited, the path pattern will be
-   *   modified as follows. Any version prefix will be replaced by the
-   *   version of the including API plus the [root][] path if specified.
-   * Example of a simple mixin:
-   *     package google.acl.v1;
-   *     service AccessControl {
-   *       // Get the underlying ACL object.
-   *       rpc GetAcl(GetAclRequest) returns (Acl) {
-   *         option (google.api.http).get = "/v1/{resource=**}:getAcl";
-   *       }
-   *     }
-   *     package google.storage.v2;
-   *     service Storage {
-   *       rpc GetAcl(GetAclRequest) returns (Acl);
-   *       // Get a data record.
-   *       rpc GetData(GetDataRequest) returns (Data) {
-   *         option (google.api.http).get = "/v2/{resource=**}";
-   *       }
-   *     }
-   * Example of a mixin configuration:
-   *     apis:
-   *     - name: google.storage.v2.Storage
-   *       mixins:
-   *       - name: google.acl.v1.AccessControl
-   * The mixin construct implies that all methods in `AccessControl` are
-   * also declared with same name and request/response types in
-   * `Storage`. A documentation generator or annotation processor will
-   * see the effective `Storage.GetAcl` method after inherting
-   * documentation and annotations as follows:
-   *     service Storage {
-   *       // Get the underlying ACL object.
-   *       rpc GetAcl(GetAclRequest) returns (Acl) {
-   *         option (google.api.http).get = "/v2/{resource=**}:getAcl";
-   *       }
-   *       ...
-   *     }
-   * Note how the version in the path pattern changed from `v1` to `v2`.
-   * If the `root` field in the mixin is specified, it should be a
-   * relative path under which inherited HTTP paths are placed. Example:
-   *     apis:
-   *     - name: google.storage.v2.Storage
-   *       mixins:
-   *       - name: google.acl.v1.AccessControl
-   *         root: acls
-   * This implies the following inherited HTTP annotation:
-   *     service Storage {
-   *       // Get the underlying ACL object.
-   *       rpc GetAcl(GetAclRequest) returns (Acl) {
-   *         option (google.api.http).get = "/v2/acls/{resource=**}:getAcl";
-   *       }
-   *       ...
-   *     }
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Mixin}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Mixin)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Mixin_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Mixin_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      name_ = "";
-
-      root_ = "";
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.ApiProto.internal_static_google_protobuf_Mixin_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin(this);
-      result.name_ = name_;
-      result.root_ = root_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin.getDefaultInstance()) return this;
-      if (!other.getName().isEmpty()) {
-        name_ = other.name_;
-        onChanged();
-      }
-      if (!other.getRoot().isEmpty()) {
-        root_ = other.root_;
-        onChanged();
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private java.lang.Object name_ = "";
-    /**
-     * <pre>
-     * The fully qualified name of the API which is included.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public java.lang.String getName() {
-      java.lang.Object ref = name_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        name_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The fully qualified name of the API which is included.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getNameBytes() {
-      java.lang.Object ref = name_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        name_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The fully qualified name of the API which is included.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      name_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The fully qualified name of the API which is included.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder clearName() {
-      
-      name_ = getDefaultInstance().getName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The fully qualified name of the API which is included.
-     * </pre>
-     *
-     * <code>string name = 1;</code>
-     */
-    public Builder setNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      name_ = value;
-      onChanged();
-      return this;
-    }
-
-    private java.lang.Object root_ = "";
-    /**
-     * <pre>
-     * If non-empty specifies a path under which inherited HTTP paths
-     * are rooted.
-     * </pre>
-     *
-     * <code>string root = 2;</code>
-     */
-    public java.lang.String getRoot() {
-      java.lang.Object ref = root_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        root_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * If non-empty specifies a path under which inherited HTTP paths
-     * are rooted.
-     * </pre>
-     *
-     * <code>string root = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getRootBytes() {
-      java.lang.Object ref = root_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        root_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * If non-empty specifies a path under which inherited HTTP paths
-     * are rooted.
-     * </pre>
-     *
-     * <code>string root = 2;</code>
-     */
-    public Builder setRoot(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      root_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * If non-empty specifies a path under which inherited HTTP paths
-     * are rooted.
-     * </pre>
-     *
-     * <code>string root = 2;</code>
-     */
-    public Builder clearRoot() {
-      
-      root_ = getDefaultInstance().getRoot();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * If non-empty specifies a path under which inherited HTTP paths
-     * are rooted.
-     * </pre>
-     *
-     * <code>string root = 2;</code>
-     */
-    public Builder setRootBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      root_ = value;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Mixin)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Mixin)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Mixin>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Mixin>() {
-    public Mixin parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Mixin(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Mixin> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Mixin> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MixinOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MixinOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MixinOrBuilder.java
deleted file mode 100644
index 99e31c6..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/MixinOrBuilder.java
+++ /dev/null
@@ -1,47 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/api.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface MixinOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Mixin)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The fully qualified name of the API which is included.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  java.lang.String getName();
-  /**
-   * <pre>
-   * The fully qualified name of the API which is included.
-   * </pre>
-   *
-   * <code>string name = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes();
-
-  /**
-   * <pre>
-   * If non-empty specifies a path under which inherited HTTP paths
-   * are rooted.
-   * </pre>
-   *
-   * <code>string root = 2;</code>
-   */
-  java.lang.String getRoot();
-  /**
-   * <pre>
-   * If non-empty specifies a path under which inherited HTTP paths
-   * are rooted.
-   * </pre>
-   *
-   * <code>string root = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getRootBytes();
-}


[23/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int32Value.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int32Value.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int32Value.java
deleted file mode 100644
index 476b086..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int32Value.java
+++ /dev/null
@@ -1,451 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `int32`.
- * The JSON representation for `Int32Value` is JSON number.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Int32Value}
- */
-public  final class Int32Value extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Int32Value)
-    Int32ValueOrBuilder {
-  // Use Int32Value.newBuilder() to construct.
-  private Int32Value(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Int32Value() {
-    value_ = 0;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Int32Value(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-
-            value_ = input.readInt32();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int32Value_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int32Value_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private int value_;
-  /**
-   * <pre>
-   * The int32 value.
-   * </pre>
-   *
-   * <code>int32 value = 1;</code>
-   */
-  public int getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (value_ != 0) {
-      output.writeInt32(1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (value_ != 0) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt32Size(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value) obj;
-
-    boolean result = true;
-    result = result && (getValue()
-        == other.getValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + getValue();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `int32`.
-   * The JSON representation for `Int32Value` is JSON number.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Int32Value}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Int32Value)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32ValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int32Value_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int32Value_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = 0;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int32Value_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value.getDefaultInstance()) return this;
-      if (other.getValue() != 0) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private int value_ ;
-    /**
-     * <pre>
-     * The int32 value.
-     * </pre>
-     *
-     * <code>int32 value = 1;</code>
-     */
-    public int getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * The int32 value.
-     * </pre>
-     *
-     * <code>int32 value = 1;</code>
-     */
-    public Builder setValue(int value) {
-      
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The int32 value.
-     * </pre>
-     *
-     * <code>int32 value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = 0;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Int32Value)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Int32Value)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Int32Value>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Int32Value>() {
-    public Int32Value parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Int32Value(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Int32Value> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Int32Value> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Int32Value getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int32ValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int32ValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int32ValueOrBuilder.java
deleted file mode 100644
index f7b0b25..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int32ValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface Int32ValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Int32Value)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The int32 value.
-   * </pre>
-   *
-   * <code>int32 value = 1;</code>
-   */
-  int getValue();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int64Value.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int64Value.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int64Value.java
deleted file mode 100644
index fce2b73..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int64Value.java
+++ /dev/null
@@ -1,452 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `int64`.
- * The JSON representation for `Int64Value` is JSON string.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Int64Value}
- */
-public  final class Int64Value extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Int64Value)
-    Int64ValueOrBuilder {
-  // Use Int64Value.newBuilder() to construct.
-  private Int64Value(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Int64Value() {
-    value_ = 0L;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Int64Value(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-
-            value_ = input.readInt64();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int64Value_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int64Value_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private long value_;
-  /**
-   * <pre>
-   * The int64 value.
-   * </pre>
-   *
-   * <code>int64 value = 1;</code>
-   */
-  public long getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (value_ != 0L) {
-      output.writeInt64(1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (value_ != 0L) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt64Size(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value) obj;
-
-    boolean result = true;
-    result = result && (getValue()
-        == other.getValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
-        getValue());
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `int64`.
-   * The JSON representation for `Int64Value` is JSON string.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Int64Value}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Int64Value)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64ValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int64Value_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int64Value_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = 0L;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_Int64Value_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value.getDefaultInstance()) return this;
-      if (other.getValue() != 0L) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private long value_ ;
-    /**
-     * <pre>
-     * The int64 value.
-     * </pre>
-     *
-     * <code>int64 value = 1;</code>
-     */
-    public long getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * The int64 value.
-     * </pre>
-     *
-     * <code>int64 value = 1;</code>
-     */
-    public Builder setValue(long value) {
-      
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The int64 value.
-     * </pre>
-     *
-     * <code>int64 value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = 0L;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Int64Value)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Int64Value)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Int64Value>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Int64Value>() {
-    public Int64Value parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Int64Value(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Int64Value> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Int64Value> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Int64Value getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int64ValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int64ValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int64ValueOrBuilder.java
deleted file mode 100644
index 291f5ed..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Int64ValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface Int64ValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Int64Value)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The int64 value.
-   * </pre>
-   *
-   * <code>int64 value = 1;</code>
-   */
-  long getValue();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/IntArrayList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/IntArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/IntArrayList.java
deleted file mode 100644
index ba8b7ec..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/IntArrayList.java
+++ /dev/null
@@ -1,272 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.IntList;
-
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.RandomAccess;
-
-/**
- * An implementation of {@link IntList} on top of a primitive array.
- *
- * @author dweis@google.com (Daniel Weis)
- */
-final class IntArrayList
-    extends AbstractProtobufList<Integer>
-    implements IntList, RandomAccess {
-
-  private static final IntArrayList EMPTY_LIST = new IntArrayList();
-  static {
-    EMPTY_LIST.makeImmutable();
-  }
-
-  public static IntArrayList emptyList() {
-    return EMPTY_LIST;
-  }
-
-  /**
-   * The backing store for the list.
-   */
-  private int[] array;
-
-  /**
-   * The size of the list distinct from the length of the array. That is, it is the number of
-   * elements set in the list.
-   */
-  private int size;
-
-  /**
-   * Constructs a new mutable {@code IntArrayList} with default capacity.
-   */
-  IntArrayList() {
-    this(new int[DEFAULT_CAPACITY], 0);
-  }
-
-  /**
-   * Constructs a new mutable {@code IntArrayList}
-   * containing the same elements as {@code other}.
-   */
-  private IntArrayList(int[] other, int size) {
-    array = other;
-    this.size = size;
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (this == o) {
-      return true;
-    }
-    if (!(o instanceof IntArrayList)) {
-      return super.equals(o);
-    }
-    IntArrayList other = (IntArrayList) o;
-    if (size != other.size) {
-      return false;
-    }
-
-    final int[] arr = other.array;
-    for (int i = 0; i < size; i++) {
-      if (array[i] != arr[i]) {
-        return false;
-      }
-    }
-
-    return true;
-  }
-
-  @Override
-  public int hashCode() {
-    int result = 1;
-    for (int i = 0; i < size; i++) {
-      result = (31 * result) + array[i];
-    }
-    return result;
-  }
-
-  @Override
-  public IntList mutableCopyWithCapacity(int capacity) {
-    if (capacity < size) {
-      throw new IllegalArgumentException();
-    }
-    return new IntArrayList(Arrays.copyOf(array, capacity), size);
-  }
-
-  @Override
-  public Integer get(int index) {
-    return getInt(index);
-  }
-
-  @Override
-  public int getInt(int index) {
-    ensureIndexInRange(index);
-    return array[index];
-  }
-
-  @Override
-  public int size() {
-    return size;
-  }
-
-  @Override
-  public Integer set(int index, Integer element) {
-    return setInt(index, element);
-  }
-
-  @Override
-  public int setInt(int index, int element) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    int previousValue = array[index];
-    array[index] = element;
-    return previousValue;
-  }
-
-  @Override
-  public void add(int index, Integer element) {
-    addInt(index, element);
-  }
-
-  /**
-   * Like {@link #add(Integer)} but more efficient in that it doesn't box the element.
-   */
-  @Override
-  public void addInt(int element) {
-    addInt(size, element);
-  }
-
-  /**
-   * Like {@link #add(int, Integer)} but more efficient in that it doesn't box the element.
-   */
-  private void addInt(int index, int element) {
-    ensureIsMutable();
-    if (index < 0 || index > size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-
-    if (size < array.length) {
-      // Shift everything over to make room
-      System.arraycopy(array, index, array, index + 1, size - index);
-    } else {
-      // Resize to 1.5x the size
-      int length = ((size * 3) / 2) + 1;
-      int[] newArray = new int[length];
-
-      // Copy the first part directly
-      System.arraycopy(array, 0, newArray, 0, index);
-
-      // Copy the rest shifted over by one to make room
-      System.arraycopy(array, index, newArray, index + 1, size - index);
-      array = newArray;
-    }
-
-    array[index] = element;
-    size++;
-    modCount++;
-  }
-
-  @Override
-  public boolean addAll(Collection<? extends Integer> collection) {
-    ensureIsMutable();
-
-    if (collection == null) {
-      throw new NullPointerException();
-    }
-
-    // We specialize when adding another IntArrayList to avoid boxing elements.
-    if (!(collection instanceof IntArrayList)) {
-      return super.addAll(collection);
-    }
-
-    IntArrayList list = (IntArrayList) collection;
-    if (list.size == 0) {
-      return false;
-    }
-
-    int overflow = Integer.MAX_VALUE - size;
-    if (overflow < list.size) {
-      // We can't actually represent a list this large.
-      throw new OutOfMemoryError();
-    }
-
-    int newSize = size + list.size;
-    if (newSize > array.length) {
-      array = Arrays.copyOf(array, newSize);
-    }
-
-    System.arraycopy(list.array, 0, array, size, list.size);
-    size = newSize;
-    modCount++;
-    return true;
-  }
-
-  @Override
-  public boolean remove(Object o) {
-    ensureIsMutable();
-    for (int i = 0; i < size; i++) {
-      if (o.equals(array[i])) {
-        System.arraycopy(array, i + 1, array, i, size - i);
-        size--;
-        modCount++;
-        return true;
-      }
-    }
-    return false;
-  }
-
-  @Override
-  public Integer remove(int index) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    int value = array[index];
-    System.arraycopy(array, index + 1, array, index, size - index);
-    size--;
-    modCount++;
-    return value;
-  }
-
-  /**
-   * Ensures that the provided {@code index} is within the range of {@code [0, size]}. Throws an
-   * {@link IndexOutOfBoundsException} if it is not.
-   *
-   * @param index the index to verify is in range
-   */
-  private void ensureIndexInRange(int index) {
-    if (index < 0 || index >= size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-  }
-
-  private String makeOutOfBoundsExceptionMessage(int index) {
-    return "Index:" + index + ", Size:" + size;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Internal.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Internal.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Internal.java
deleted file mode 100644
index 8f5c229..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Internal.java
+++ /dev/null
@@ -1,751 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-import java.lang.reflect.Method;
-import java.nio.ByteBuffer;
-import java.nio.charset.Charset;
-import java.util.AbstractList;
-import java.util.AbstractMap;
-import java.util.AbstractSet;
-import java.util.Arrays;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-import java.util.RandomAccess;
-import java.util.Set;
-
-/**
- * The classes contained within are used internally by the Protocol Buffer
- * library and generated message implementations. They are public only because
- * those generated messages do not reside in the {@code protobuf} package.
- * Others should not use this class directly.
- *
- * @author kenton@google.com (Kenton Varda)
- */
-public final class Internal {
-
-  private Internal() {}
-
-  static final Charset UTF_8 = Charset.forName("UTF-8");
-  static final Charset ISO_8859_1 = Charset.forName("ISO-8859-1");
-
-  /**
-   * Throws an appropriate {@link NullPointerException} if the given objects is {@code null}.
-   */
-  static <T> T checkNotNull(T obj, String message) {
-    if (obj == null) {
-      throw new NullPointerException(message);
-    }
-    return obj;
-  }
-
-  /**
-   * Helper called by generated code to construct default values for string
-   * fields.
-   * <p>
-   * The protocol compiler does not actually contain a UTF-8 decoder -- it
-   * just pushes UTF-8-encoded text around without touching it.  The one place
-   * where this presents a problem is when generating Java string literals.
-   * Unicode characters in the string literal would normally need to be encoded
-   * using a Unicode escape sequence, which would require decoding them.
-   * To get around this, protoc instead embeds the UTF-8 bytes into the
-   * generated code and leaves it to the runtime library to decode them.
-   * <p>
-   * It gets worse, though.  If protoc just generated a byte array, like:
-   *   new byte[] {0x12, 0x34, 0x56, 0x78}
-   * Java actually generates *code* which allocates an array and then fills
-   * in each value.  This is much less efficient than just embedding the bytes
-   * directly into the bytecode.  To get around this, we need another
-   * work-around.  String literals are embedded directly, so protoc actually
-   * generates a string literal corresponding to the bytes.  The easiest way
-   * to do this is to use the ISO-8859-1 character set, which corresponds to
-   * the first 256 characters of the Unicode range.  Protoc can then use
-   * good old CEscape to generate the string.
-   * <p>
-   * So we have a string literal which represents a set of bytes which
-   * represents another string.  This function -- stringDefaultValue --
-   * converts from the generated string to the string we actually want.  The
-   * generated code calls this automatically.
-   */
-  public static String stringDefaultValue(String bytes) {
-    return new String(bytes.getBytes(ISO_8859_1), UTF_8);
-  }
-
-  /**
-   * Helper called by generated code to construct default values for bytes
-   * fields.
-   * <p>
-   * This is a lot like {@link #stringDefaultValue}, but for bytes fields.
-   * In this case we only need the second of the two hacks -- allowing us to
-   * embed raw bytes as a string literal with ISO-8859-1 encoding.
-   */
-  public static ByteString bytesDefaultValue(String bytes) {
-    return ByteString.copyFrom(bytes.getBytes(ISO_8859_1));
-  }
-  /**
-   * Helper called by generated code to construct default values for bytes
-   * fields.
-   * <p>
-   * This is like {@link #bytesDefaultValue}, but returns a byte array.
-   */
-  public static byte[] byteArrayDefaultValue(String bytes) {
-    return bytes.getBytes(ISO_8859_1);
-  }
-
-  /**
-   * Helper called by generated code to construct default values for bytes
-   * fields.
-   * <p>
-   * This is like {@link #bytesDefaultValue}, but returns a ByteBuffer.
-   */
-  public static ByteBuffer byteBufferDefaultValue(String bytes) {
-    return ByteBuffer.wrap(byteArrayDefaultValue(bytes));
-  }
-
-  /**
-   * Create a new ByteBuffer and copy all the content of {@code source}
-   * ByteBuffer to the new ByteBuffer. The new ByteBuffer's limit and
-   * capacity will be source.capacity(), and its position will be 0.
-   * Note that the state of {@code source} ByteBuffer won't be changed.
-   */
-  public static ByteBuffer copyByteBuffer(ByteBuffer source) {
-    // Make a duplicate of the source ByteBuffer and read data from the
-    // duplicate. This is to avoid affecting the source ByteBuffer's state.
-    ByteBuffer temp = source.duplicate();
-    // We want to copy all the data in the source ByteBuffer, not just the
-    // remaining bytes.
-    temp.clear();
-    ByteBuffer result = ByteBuffer.allocate(temp.capacity());
-    result.put(temp);
-    result.clear();
-    return result;
-  }
-
-  /**
-   * Helper called by generated code to determine if a byte array is a valid
-   * UTF-8 encoded string such that the original bytes can be converted to
-   * a String object and then back to a byte array round tripping the bytes
-   * without loss.  More precisely, returns {@code true} whenever:
-   * <pre>   {@code
-   * Arrays.equals(byteString.toByteArray(),
-   *     new String(byteString.toByteArray(), "UTF-8").getBytes("UTF-8"))
-   * }</pre>
-   *
-   * <p>This method rejects "overlong" byte sequences, as well as
-   * 3-byte sequences that would map to a surrogate character, in
-   * accordance with the restricted definition of UTF-8 introduced in
-   * Unicode 3.1.  Note that the UTF-8 decoder included in Oracle's
-   * JDK has been modified to also reject "overlong" byte sequences,
-   * but currently (2011) still accepts 3-byte surrogate character
-   * byte sequences.
-   *
-   * <p>See the Unicode Standard,<br>
-   * Table 3-6. <em>UTF-8 Bit Distribution</em>,<br>
-   * Table 3-7. <em>Well Formed UTF-8 Byte Sequences</em>.
-   *
-   * <p>As of 2011-02, this method simply returns the result of {@link
-   * ByteString#isValidUtf8()}.  Calling that method directly is preferred.
-   *
-   * @param byteString the string to check
-   * @return whether the byte array is round trippable
-   */
-  public static boolean isValidUtf8(ByteString byteString) {
-    return byteString.isValidUtf8();
-  }
-
-  /**
-   * Like {@link #isValidUtf8(ByteString)} but for byte arrays.
-   */
-  public static boolean isValidUtf8(byte[] byteArray) {
-    return Utf8.isValidUtf8(byteArray);
-  }
-
-  /**
-   * Helper method to get the UTF-8 bytes of a string.
-   */
-  public static byte[] toByteArray(String value) {
-    return value.getBytes(UTF_8);
-  }
-
-  /**
-   * Helper method to convert a byte array to a string using UTF-8 encoding.
-   */
-  public static String toStringUtf8(byte[] bytes) {
-    return new String(bytes, UTF_8);
-  }
-
-  /**
-   * Interface for an enum value or value descriptor, to be used in FieldSet.
-   * The lite library stores enum values directly in FieldSets but the full
-   * library stores EnumValueDescriptors in order to better support reflection.
-   */
-  public interface EnumLite {
-    int getNumber();
-  }
-
-  /**
-   * Interface for an object which maps integers to {@link EnumLite}s.
-   * {@link Descriptors.EnumDescriptor} implements this interface by mapping
-   * numbers to {@link Descriptors.EnumValueDescriptor}s.  Additionally,
-   * every generated enum type has a static method internalGetValueMap() which
-   * returns an implementation of this type that maps numbers to enum values.
-   */
-  public interface EnumLiteMap<T extends EnumLite> {
-    T findValueByNumber(int number);
-  }
-
-  /**
-   * Helper method for implementing {@link Message#hashCode()} for longs.
-   * @see Long#hashCode()
-   */
-  public static int hashLong(long n) {
-    return (int) (n ^ (n >>> 32));
-  }
-
-  /**
-   * Helper method for implementing {@link Message#hashCode()} for
-   * booleans.
-   * @see Boolean#hashCode()
-   */
-  public static int hashBoolean(boolean b) {
-    return b ? 1231 : 1237;
-  }
-
-  /**
-   * Helper method for implementing {@link Message#hashCode()} for enums.
-   * <p>
-   * This is needed because {@link java.lang.Enum#hashCode()} is final, but we
-   * need to use the field number as the hash code to ensure compatibility
-   * between statically and dynamically generated enum objects.
-   */
-  public static int hashEnum(EnumLite e) {
-    return e.getNumber();
-  }
-
-  /**
-   * Helper method for implementing {@link Message#hashCode()} for
-   * enum lists.
-   */
-  public static int hashEnumList(List<? extends EnumLite> list) {
-    int hash = 1;
-    for (EnumLite e : list) {
-      hash = 31 * hash + hashEnum(e);
-    }
-    return hash;
-  }
-
-  /**
-   * Helper method for implementing {@link Message#equals(Object)} for bytes field.
-   */
-  public static boolean equals(List<byte[]> a, List<byte[]> b) {
-    if (a.size() != b.size()) return false;
-    for (int i = 0; i < a.size(); ++i) {
-      if (!Arrays.equals(a.get(i), b.get(i))) {
-        return false;
-      }
-    }
-    return true;
-  }
-
-  /**
-   * Helper method for implementing {@link Message#hashCode()} for bytes field.
-   */
-  public static int hashCode(List<byte[]> list) {
-    int hash = 1;
-    for (byte[] bytes : list) {
-      hash = 31 * hash + hashCode(bytes);
-    }
-    return hash;
-  }
-
-  /**
-   * Helper method for implementing {@link Message#hashCode()} for bytes field.
-   */
-  public static int hashCode(byte[] bytes) {
-    // The hash code for a byte array should be the same as the hash code for a
-    // ByteString with the same content. This is to ensure that the generated
-    // hashCode() method will return the same value as the pure reflection
-    // based hashCode() method.
-    return Internal.hashCode(bytes, 0, bytes.length);
-  }
-  
-  /**
-   * Helper method for implementing {@link LiteralByteString#hashCode()}.
-   */
-  static int hashCode(byte[] bytes, int offset, int length) {
-    // The hash code for a byte array should be the same as the hash code for a
-    // ByteString with the same content. This is to ensure that the generated
-    // hashCode() method will return the same value as the pure reflection
-    // based hashCode() method.
-    int h = Internal.partialHash(length, bytes, offset, length);
-    return h == 0 ? 1 : h;
-  }
-
-  /**
-   * Helper method for continuously hashing bytes.
-   */
-  static int partialHash(int h, byte[] bytes, int offset, int length) {
-    for (int i = offset; i < offset + length; i++) {
-      h = h * 31 + bytes[i];
-    }
-    return h;
-  }
-  
-  /**
-   * Helper method for implementing {@link Message#equals(Object)} for bytes
-   * field.
-   */
-  public static boolean equalsByteBuffer(ByteBuffer a, ByteBuffer b) {
-    if (a.capacity() != b.capacity()) {
-      return false;
-    }
-    // ByteBuffer.equals() will only compare the remaining bytes, but we want to
-    // compare all the content.
-    return a.duplicate().clear().equals(b.duplicate().clear());
-  }
-
-  /**
-   * Helper method for implementing {@link Message#equals(Object)} for bytes
-   * field.
-   */
-  public static boolean equalsByteBuffer(
-      List<ByteBuffer> a, List<ByteBuffer> b) {
-    if (a.size() != b.size()) {
-      return false;
-    }
-    for (int i = 0; i < a.size(); ++i) {
-      if (!equalsByteBuffer(a.get(i), b.get(i))) {
-        return false;
-      }
-    }
-    return true;
-  }
-
-  /**
-   * Helper method for implementing {@link Message#hashCode()} for bytes
-   * field.
-   */
-  public static int hashCodeByteBuffer(List<ByteBuffer> list) {
-    int hash = 1;
-    for (ByteBuffer bytes : list) {
-      hash = 31 * hash + hashCodeByteBuffer(bytes);
-    }
-    return hash;
-  }
-
-  private static final int DEFAULT_BUFFER_SIZE = 4096;
-
-  /**
-   * Helper method for implementing {@link Message#hashCode()} for bytes
-   * field.
-   */
-  public static int hashCodeByteBuffer(ByteBuffer bytes) {
-    if (bytes.hasArray()) {
-      // Fast path.
-      int h = partialHash(bytes.capacity(), bytes.array(), bytes.arrayOffset(), bytes.capacity());
-      return h == 0 ? 1 : h;
-    } else {
-      // Read the data into a temporary byte array before calculating the
-      // hash value.
-      final int bufferSize = bytes.capacity() > DEFAULT_BUFFER_SIZE
-          ? DEFAULT_BUFFER_SIZE : bytes.capacity();
-      final byte[] buffer = new byte[bufferSize];
-      final ByteBuffer duplicated = bytes.duplicate();
-      duplicated.clear();
-      int h = bytes.capacity();
-      while (duplicated.remaining() > 0) {
-        final int length = duplicated.remaining() <= bufferSize ?
-            duplicated.remaining() : bufferSize;
-        duplicated.get(buffer, 0, length);
-        h = partialHash(h, buffer, 0, length);
-      }
-      return h == 0 ? 1 : h;
-    }
-  }
-
-  @SuppressWarnings("unchecked")
-  public static <T extends MessageLite> T getDefaultInstance(Class<T> clazz) {
-    try {
-      Method method = clazz.getMethod("getDefaultInstance");
-      return (T) method.invoke(method);
-    } catch (Exception e) {
-      throw new RuntimeException(
-          "Failed to get default instance for " + clazz, e);
-    }
-  }
-
-  /**
-   * An empty byte array constant used in generated code.
-   */
-  public static final byte[] EMPTY_BYTE_ARRAY = new byte[0];
-
-  /**
-   * An empty byte array constant used in generated code.
-   */
-  public static final ByteBuffer EMPTY_BYTE_BUFFER =
-      ByteBuffer.wrap(EMPTY_BYTE_ARRAY);
-
-  /** An empty coded input stream constant used in generated code. */
-  public static final CodedInputStream EMPTY_CODED_INPUT_STREAM =
-      CodedInputStream.newInstance(EMPTY_BYTE_ARRAY);
-
-
-  /**
-   * Provides an immutable view of {@code List<T>} around a {@code List<F>}.
-   *
-   * Protobuf internal. Used in protobuf generated code only.
-   */
-  public static class ListAdapter<F, T> extends AbstractList<T> {
-    /**
-     * Convert individual elements of the List from F to T.
-     */
-    public interface Converter<F, T> {
-      T convert(F from);
-    }
-
-    private final List<F> fromList;
-    private final Converter<F, T> converter;
-
-    public ListAdapter(List<F> fromList, Converter<F, T> converter) {
-      this.fromList = fromList;
-      this.converter = converter;
-    }
-
-    @Override
-    public T get(int index) {
-      return converter.convert(fromList.get(index));
-    }
-
-    @Override
-    public int size() {
-      return fromList.size();
-    }
-  }
-
-  /**
-   * Wrap around a {@code Map<K, RealValue>} and provide a {@code Map<K, V>}
-   * interface.
-   */
-  public static class MapAdapter<K, V, RealValue> extends AbstractMap<K, V> {
-    /**
-     * An interface used to convert between two types.
-     */
-    public interface Converter<A, B> {
-      B doForward(A object);
-      A doBackward(B object);
-    }
-
-    public static <T extends EnumLite> Converter<Integer, T> newEnumConverter(
-        final EnumLiteMap<T> enumMap, final T unrecognizedValue) {
-      return new Converter<Integer, T>() {
-        @Override
-        public T doForward(Integer value) {
-          T result = enumMap.findValueByNumber(value);
-          return result == null ? unrecognizedValue : result;
-        }
-
-        @Override
-        public Integer doBackward(T value) {
-          return value.getNumber();
-        }
-      };
-    }
-
-    private final Map<K, RealValue> realMap;
-    private final Converter<RealValue, V> valueConverter;
-
-    public MapAdapter(Map<K, RealValue> realMap,
-        Converter<RealValue, V> valueConverter) {
-      this.realMap = realMap;
-      this.valueConverter = valueConverter;
-    }
-
-    @SuppressWarnings("unchecked")
-    @Override
-    public V get(Object key) {
-      RealValue result = realMap.get(key);
-      if (result == null) {
-        return null;
-      }
-      return valueConverter.doForward(result);
-    }
-
-    @Override
-    public V put(K key, V value) {
-      RealValue oldValue = realMap.put(key, valueConverter.doBackward(value));
-      if (oldValue == null) {
-        return null;
-      }
-      return valueConverter.doForward(oldValue);
-    }
-
-    @Override
-    public Set<java.util.Map.Entry<K, V>> entrySet() {
-      return new SetAdapter(realMap.entrySet());
-    }
-
-    private class SetAdapter extends AbstractSet<Map.Entry<K, V>> {
-      private final Set<Map.Entry<K, RealValue>> realSet;
-      public SetAdapter(Set<Map.Entry<K, RealValue>> realSet) {
-        this.realSet = realSet;
-      }
-
-      @Override
-      public Iterator<java.util.Map.Entry<K, V>> iterator() {
-        return new IteratorAdapter(realSet.iterator());
-      }
-
-      @Override
-      public int size() {
-        return realSet.size();
-      }
-    }
-
-    private class IteratorAdapter implements Iterator<Map.Entry<K, V>> {
-      private final Iterator<Map.Entry<K, RealValue>> realIterator;
-
-      public IteratorAdapter(
-          Iterator<Map.Entry<K, RealValue>> realIterator) {
-        this.realIterator = realIterator;
-      }
-
-      @Override
-      public boolean hasNext() {
-        return realIterator.hasNext();
-      }
-
-      @Override
-      public java.util.Map.Entry<K, V> next() {
-        return new EntryAdapter(realIterator.next());
-      }
-
-      @Override
-      public void remove() {
-        realIterator.remove();
-      }
-    }
-
-    private class EntryAdapter implements Map.Entry<K, V> {
-      private final Map.Entry<K, RealValue> realEntry;
-
-      public EntryAdapter(Map.Entry<K, RealValue> realEntry) {
-        this.realEntry = realEntry;
-      }
-
-      @Override
-      public K getKey() {
-        return realEntry.getKey();
-      }
-
-      @Override
-      public V getValue() {
-        return valueConverter.doForward(realEntry.getValue());
-      }
-
-      @Override
-      public V setValue(V value) {
-        RealValue oldValue = realEntry.setValue(
-            valueConverter.doBackward(value));
-        if (oldValue == null) {
-          return null;
-        }
-        return valueConverter.doForward(oldValue);
-      }
-    }
-  }
-
-  /**
-   * Extends {@link List} to add the capability to make the list immutable and inspect if it is
-   * modifiable.
-   * <p>
-   * All implementations must support efficient random access.
-   */
-  public static interface ProtobufList<E> extends List<E>, RandomAccess {
-
-    /**
-     * Makes this list immutable. All subsequent modifications will throw an
-     * {@link UnsupportedOperationException}.
-     */
-    void makeImmutable();
-
-    /**
-     * Returns whether this list can be modified via the publicly accessible {@link List} methods.
-     */
-    boolean isModifiable();
-
-    /**
-     * Returns a mutable clone of this list with the specified capacity.
-     */
-    ProtobufList<E> mutableCopyWithCapacity(int capacity);
-  }
-
-  /**
-   * A {@link java.util.List} implementation that avoids boxing the elements into Integers if
-   * possible. Does not support null elements.
-   */
-  public static interface IntList extends ProtobufList<Integer> {
-
-    /**
-     * Like {@link #get(int)} but more efficient in that it doesn't box the returned value.
-     */
-    int getInt(int index);
-
-    /**
-     * Like {@link #add(Object)} but more efficient in that it doesn't box the element.
-     */
-    void addInt(int element);
-
-    /**
-     * Like {@link #set(int, Object)} but more efficient in that it doesn't box the element.
-     */
-    int setInt(int index, int element);
-
-    /**
-     * Returns a mutable clone of this list with the specified capacity.
-     */
-    @Override
-    IntList mutableCopyWithCapacity(int capacity);
-  }
-
-  /**
-   * A {@link java.util.List} implementation that avoids boxing the elements into Booleans if
-   * possible. Does not support null elements.
-   */
-  public static interface BooleanList extends ProtobufList<Boolean> {
-
-    /**
-     * Like {@link #get(int)} but more efficient in that it doesn't box the returned value.
-     */
-    boolean getBoolean(int index);
-
-    /**
-     * Like {@link #add(Object)} but more efficient in that it doesn't box the element.
-     */
-    void addBoolean(boolean element);
-
-    /**
-     * Like {@link #set(int, Object)} but more efficient in that it doesn't box the element.
-     */
-    boolean setBoolean(int index, boolean element);
-
-    /**
-     * Returns a mutable clone of this list with the specified capacity.
-     */
-    @Override
-    BooleanList mutableCopyWithCapacity(int capacity);
-  }
-
-  /**
-   * A {@link java.util.List} implementation that avoids boxing the elements into Longs if
-   * possible. Does not support null elements.
-   */
-  public static interface LongList extends ProtobufList<Long> {
-
-    /**
-     * Like {@link #get(int)} but more efficient in that it doesn't box the returned value.
-     */
-    long getLong(int index);
-
-    /**
-     * Like {@link #add(Object)} but more efficient in that it doesn't box the element.
-     */
-    void addLong(long element);
-
-    /**
-     * Like {@link #set(int, Object)} but more efficient in that it doesn't box the element.
-     */
-    long setLong(int index, long element);
-
-    /**
-     * Returns a mutable clone of this list with the specified capacity.
-     */
-    @Override
-    LongList mutableCopyWithCapacity(int capacity);
-  }
-
-  /**
-   * A {@link java.util.List} implementation that avoids boxing the elements into Doubles if
-   * possible. Does not support null elements.
-   */
-  public static interface DoubleList extends ProtobufList<Double> {
-
-    /**
-     * Like {@link #get(int)} but more efficient in that it doesn't box the returned value.
-     */
-    double getDouble(int index);
-
-    /**
-     * Like {@link #add(Object)} but more efficient in that it doesn't box the element.
-     */
-    void addDouble(double element);
-
-    /**
-     * Like {@link #set(int, Object)} but more efficient in that it doesn't box the element.
-     */
-    double setDouble(int index, double element);
-
-    /**
-     * Returns a mutable clone of this list with the specified capacity.
-     */
-    @Override
-    DoubleList mutableCopyWithCapacity(int capacity);
-  }
-
-  /**
-   * A {@link java.util.List} implementation that avoids boxing the elements into Floats if
-   * possible. Does not support null elements.
-   */
-  public static interface FloatList extends ProtobufList<Float> {
-
-    /**
-     * Like {@link #get(int)} but more efficient in that it doesn't box the returned value.
-     */
-    float getFloat(int index);
-
-    /**
-     * Like {@link #add(Object)} but more efficient in that it doesn't box the element.
-     */
-    void addFloat(float element);
-
-    /**
-     * Like {@link #set(int, Object)} but more efficient in that it doesn't box the element.
-     */
-    float setFloat(int index, float element);
-
-    /**
-     * Returns a mutable clone of this list with the specified capacity.
-     */
-    @Override
-    FloatList mutableCopyWithCapacity(int capacity);
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/InvalidProtocolBufferException.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/InvalidProtocolBufferException.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/InvalidProtocolBufferException.java
deleted file mode 100644
index 2682242..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/InvalidProtocolBufferException.java
+++ /dev/null
@@ -1,146 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.io.IOException;
-
-/**
- * Thrown when a protocol message being parsed is invalid in some way,
- * e.g. it contains a malformed varint or a negative byte length.
- *
- * @author kenton@google.com Kenton Varda
- */
-public class InvalidProtocolBufferException extends IOException {
-  private static final long serialVersionUID = -1616151763072450476L;
-  private MessageLite unfinishedMessage = null;
-
-  public InvalidProtocolBufferException(final String description) {
-    super(description);
-  }
-
-  public InvalidProtocolBufferException(IOException e) {
-    super(e.getMessage(), e);
-  }
-
-  /**
-   * Attaches an unfinished message to the exception to support best-effort
-   * parsing in {@code Parser} interface.
-   *
-   * @return this
-   */
-  public InvalidProtocolBufferException setUnfinishedMessage(
-      MessageLite unfinishedMessage) {
-    this.unfinishedMessage = unfinishedMessage;
-    return this;
-  }
-
-  /**
-   * Returns the unfinished message attached to the exception, or null if
-   * no message is attached.
-   */
-  public MessageLite getUnfinishedMessage() {
-    return unfinishedMessage;
-  }
-
-  /**
-   * Unwraps the underlying {@link IOException} if this exception was caused by an I/O
-   * problem. Otherwise, returns {@code this}.
-   */
-  public IOException unwrapIOException() {
-    return getCause() instanceof IOException ? (IOException) getCause() : this;
-  }
-
-  static InvalidProtocolBufferException truncatedMessage() {
-    return new InvalidProtocolBufferException(
-      "While parsing a protocol message, the input ended unexpectedly " +
-      "in the middle of a field.  This could mean either that the " +
-      "input has been truncated or that an embedded message " +
-      "misreported its own length.");
-  }
-
-  static InvalidProtocolBufferException negativeSize() {
-    return new InvalidProtocolBufferException(
-      "CodedInputStream encountered an embedded string or message " +
-      "which claimed to have negative size.");
-  }
-
-  static InvalidProtocolBufferException malformedVarint() {
-    return new InvalidProtocolBufferException(
-      "CodedInputStream encountered a malformed varint.");
-  }
-
-  static InvalidProtocolBufferException invalidTag() {
-    return new InvalidProtocolBufferException(
-      "Protocol message contained an invalid tag (zero).");
-  }
-
-  static InvalidProtocolBufferException invalidEndTag() {
-    return new InvalidProtocolBufferException(
-      "Protocol message end-group tag did not match expected tag.");
-  }
-
-  static InvalidWireTypeException invalidWireType() {
-    return new InvalidWireTypeException(
-      "Protocol message tag had invalid wire type.");
-  }
-
-  /**
-   * Exception indicating that and unexpected wire type was encountered for a field.
-   */
-  @ExperimentalApi
-  public static class InvalidWireTypeException extends InvalidProtocolBufferException {
-    private static final long serialVersionUID = 3283890091615336259L;
-
-    public InvalidWireTypeException(String description) {
-      super(description);
-    }
-  }
-
-  static InvalidProtocolBufferException recursionLimitExceeded() {
-    return new InvalidProtocolBufferException(
-      "Protocol message had too many levels of nesting.  May be malicious.  " +
-      "Use CodedInputStream.setRecursionLimit() to increase the depth limit.");
-  }
-
-  static InvalidProtocolBufferException sizeLimitExceeded() {
-    return new InvalidProtocolBufferException(
-      "Protocol message was too large.  May be malicious.  " +
-      "Use CodedInputStream.setSizeLimit() to increase the size limit.");
-  }
-
-  static InvalidProtocolBufferException parseFailure() {
-    return new InvalidProtocolBufferException("Failed to parse the message.");
-  }
-
-  static InvalidProtocolBufferException invalidUtf8() {
-    return new InvalidProtocolBufferException("Protocol message had invalid UTF-8.");
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyField.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyField.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyField.java
deleted file mode 100644
index 886d596..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/LazyField.java
+++ /dev/null
@@ -1,154 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.Iterator;
-import java.util.Map.Entry;
-
-/**
- * LazyField encapsulates the logic of lazily parsing message fields. It stores
- * the message in a ByteString initially and then parse it on-demand.
- *
- * Most of key methods are implemented in {@link LazyFieldLite} but this class
- * can contain default instance of the message to provide {@code hashCode()},
- * {@code euqals()} and {@code toString()}.
- *
- * @author xiangl@google.com (Xiang Li)
- */
-public class LazyField extends LazyFieldLite {
-
-  /**
-   * Carry a message's default instance which is used by {@code hashCode()}, {@code euqals()} and
-   * {@code toString()}.
-   */
-  private final MessageLite defaultInstance;
-
-  public LazyField(MessageLite defaultInstance,
-      ExtensionRegistryLite extensionRegistry, ByteString bytes) {
-    super(extensionRegistry, bytes);
-
-    this.defaultInstance = defaultInstance;
-  }
-
-  @Override
-  public boolean containsDefaultInstance() {
-    return super.containsDefaultInstance() || value == defaultInstance;
-  }
-
-  public MessageLite getValue() {
-    return getValue(defaultInstance);
-  }
-
-  @Override
-  public int hashCode() {
-    return getValue().hashCode();
-  }
-
-  @Override
-  public boolean equals(Object obj) {
-    return getValue().equals(obj);
-  }
-
-  @Override
-  public String toString() {
-    return getValue().toString();
-  }
-
-  // ====================================================
-
-  /**
-   * LazyEntry and LazyIterator are used to encapsulate the LazyField, when
-   * users iterate all fields from FieldSet.
-   */
-  static class LazyEntry<K> implements Entry<K, Object> {
-    private Entry<K, LazyField> entry;
-
-    private LazyEntry(Entry<K, LazyField> entry) {
-      this.entry = entry;
-    }
-
-    @Override
-    public K getKey() {
-      return entry.getKey();
-    }
-
-    @Override
-    public Object getValue() {
-      LazyField field = entry.getValue();
-      if (field == null) {
-        return null;
-      }
-      return field.getValue();
-    }
-
-    public LazyField getField() {
-      return entry.getValue();
-    }
-
-    @Override
-    public Object setValue(Object value) {
-      if (!(value instanceof MessageLite)) {
-        throw new IllegalArgumentException(
-            "LazyField now only used for MessageSet, "
-            + "and the value of MessageSet must be an instance of MessageLite");
-      }
-      return entry.getValue().setValue((MessageLite) value);
-    }
-  }
-
-  static class LazyIterator<K> implements Iterator<Entry<K, Object>> {
-    private Iterator<Entry<K, Object>> iterator;
-
-    public LazyIterator(Iterator<Entry<K, Object>> iterator) {
-      this.iterator = iterator;
-    }
-
-    @Override
-    public boolean hasNext() {
-      return iterator.hasNext();
-    }
-
-    @Override
-    @SuppressWarnings("unchecked")
-    public Entry<K, Object> next() {
-      Entry<K, ?> entry = iterator.next();
-      if (entry.getValue() instanceof LazyField) {
-        return new LazyEntry<K>((Entry<K, LazyField>) entry);
-      }
-      return (Entry<K, Object>) entry;
-    }
-
-    @Override
-    public void remove() {
-      iterator.remove();
-    }
-  }
-}


[46/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/DummyRegionServerEndpointProtos.java
----------------------------------------------------------------------
diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/DummyRegionServerEndpointProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/DummyRegionServerEndpointProtos.java
deleted file mode 100644
index a011b30..0000000
--- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/DummyRegionServerEndpointProtos.java
+++ /dev/null
@@ -1,1225 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: DummyRegionServerEndpoint.proto
-
-package org.apache.hadoop.hbase.coprocessor.protobuf.generated;
-
-public final class DummyRegionServerEndpointProtos {
-  private DummyRegionServerEndpointProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface DummyRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-  }
-  /**
-   * Protobuf type {@code hbase.test.pb.DummyRequest}
-   */
-  public static final class DummyRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements DummyRequestOrBuilder {
-    // Use DummyRequest.newBuilder() to construct.
-    private DummyRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private DummyRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final DummyRequest defaultInstance;
-    public static DummyRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public DummyRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private DummyRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<DummyRequest> PARSER =
-        new com.google.protobuf.AbstractParser<DummyRequest>() {
-      public DummyRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new DummyRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<DummyRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private void initFields() {
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest) obj;
-
-      boolean result = true;
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.test.pb.DummyRequest}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest(this);
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.getDefaultInstance()) return this;
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.test.pb.DummyRequest)
-    }
-
-    static {
-      defaultInstance = new DummyRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.test.pb.DummyRequest)
-  }
-
-  public interface DummyResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required string value = 1;
-    /**
-     * <code>required string value = 1;</code>
-     */
-    boolean hasValue();
-    /**
-     * <code>required string value = 1;</code>
-     */
-    java.lang.String getValue();
-    /**
-     * <code>required string value = 1;</code>
-     */
-    com.google.protobuf.ByteString
-        getValueBytes();
-  }
-  /**
-   * Protobuf type {@code hbase.test.pb.DummyResponse}
-   */
-  public static final class DummyResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements DummyResponseOrBuilder {
-    // Use DummyResponse.newBuilder() to construct.
-    private DummyResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private DummyResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final DummyResponse defaultInstance;
-    public static DummyResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public DummyResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private DummyResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              value_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<DummyResponse> PARSER =
-        new com.google.protobuf.AbstractParser<DummyResponse>() {
-      public DummyResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new DummyResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<DummyResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required string value = 1;
-    public static final int VALUE_FIELD_NUMBER = 1;
-    private java.lang.Object value_;
-    /**
-     * <code>required string value = 1;</code>
-     */
-    public boolean hasValue() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required string value = 1;</code>
-     */
-    public java.lang.String getValue() {
-      java.lang.Object ref = value_;
-      if (ref instanceof java.lang.String) {
-        return (java.lang.String) ref;
-      } else {
-        com.google.protobuf.ByteString bs = 
-            (com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        if (bs.isValidUtf8()) {
-          value_ = s;
-        }
-        return s;
-      }
-    }
-    /**
-     * <code>required string value = 1;</code>
-     */
-    public com.google.protobuf.ByteString
-        getValueBytes() {
-      java.lang.Object ref = value_;
-      if (ref instanceof java.lang.String) {
-        com.google.protobuf.ByteString b = 
-            com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        value_ = b;
-        return b;
-      } else {
-        return (com.google.protobuf.ByteString) ref;
-      }
-    }
-
-    private void initFields() {
-      value_ = "";
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasValue()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, getValueBytes());
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, getValueBytes());
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) obj;
-
-      boolean result = true;
-      result = result && (hasValue() == other.hasValue());
-      if (hasValue()) {
-        result = result && getValue()
-            .equals(other.getValue());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasValue()) {
-        hash = (37 * hash) + VALUE_FIELD_NUMBER;
-        hash = (53 * hash) + getValue().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.test.pb.DummyResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        value_ = "";
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.internal_static_hbase_test_pb_DummyResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.value_ = value_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance()) return this;
-        if (other.hasValue()) {
-          bitField0_ |= 0x00000001;
-          value_ = other.value_;
-          onChanged();
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasValue()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required string value = 1;
-      private java.lang.Object value_ = "";
-      /**
-       * <code>required string value = 1;</code>
-       */
-      public boolean hasValue() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required string value = 1;</code>
-       */
-      public java.lang.String getValue() {
-        java.lang.Object ref = value_;
-        if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          value_ = s;
-          return s;
-        } else {
-          return (java.lang.String) ref;
-        }
-      }
-      /**
-       * <code>required string value = 1;</code>
-       */
-      public com.google.protobuf.ByteString
-          getValueBytes() {
-        java.lang.Object ref = value_;
-        if (ref instanceof String) {
-          com.google.protobuf.ByteString b = 
-              com.google.protobuf.ByteString.copyFromUtf8(
-                  (java.lang.String) ref);
-          value_ = b;
-          return b;
-        } else {
-          return (com.google.protobuf.ByteString) ref;
-        }
-      }
-      /**
-       * <code>required string value = 1;</code>
-       */
-      public Builder setValue(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        value_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string value = 1;</code>
-       */
-      public Builder clearValue() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        value_ = getDefaultInstance().getValue();
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required string value = 1;</code>
-       */
-      public Builder setValueBytes(
-          com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        value_ = value;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.test.pb.DummyResponse)
-    }
-
-    static {
-      defaultInstance = new DummyResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.test.pb.DummyResponse)
-  }
-
-  /**
-   * Protobuf service {@code hbase.test.pb.DummyService}
-   */
-  public static abstract class DummyService
-      implements com.google.protobuf.Service {
-    protected DummyService() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc dummyCall(.hbase.test.pb.DummyRequest) returns (.hbase.test.pb.DummyResponse);</code>
-       */
-      public abstract void dummyCall(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse> done);
-
-      /**
-       * <code>rpc dummyThrow(.hbase.test.pb.DummyRequest) returns (.hbase.test.pb.DummyResponse);</code>
-       */
-      public abstract void dummyThrow(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse> done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new DummyService() {
-        @java.lang.Override
-        public  void dummyCall(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse> done) {
-          impl.dummyCall(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void dummyThrow(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse> done) {
-          impl.dummyThrow(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.dummyCall(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)request);
-            case 1:
-              return impl.dummyThrow(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc dummyCall(.hbase.test.pb.DummyRequest) returns (.hbase.test.pb.DummyResponse);</code>
-     */
-    public abstract void dummyCall(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse> done);
-
-    /**
-     * <code>rpc dummyThrow(.hbase.test.pb.DummyRequest) returns (.hbase.test.pb.DummyResponse);</code>
-     */
-    public abstract void dummyThrow(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse> done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.dummyCall(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse>specializeCallback(
-              done));
-          return;
-        case 1:
-          this.dummyThrow(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyService implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void dummyCall(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.class,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance()));
-      }
-
-      public  void dummyThrow(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.class,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse dummyCall(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request)
-          throws com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse dummyThrow(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request)
-          throws com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse dummyCall(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse dummyThrow(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.DummyRegionServerEndpointProtos.DummyResponse.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.test.pb.DummyService)
-  }
-
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_hbase_test_pb_DummyRequest_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_hbase_test_pb_DummyRequest_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_hbase_test_pb_DummyResponse_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_hbase_test_pb_DummyResponse_fieldAccessorTable;
-
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\037DummyRegionServerEndpoint.proto\022\rhbase" +
-      ".test.pb\"\016\n\014DummyRequest\"\036\n\rDummyRespons" +
-      "e\022\r\n\005value\030\001 \002(\t2\237\001\n\014DummyService\022F\n\tdum" +
-      "myCall\022\033.hbase.test.pb.DummyRequest\032\034.hb" +
-      "ase.test.pb.DummyResponse\022G\n\ndummyThrow\022" +
-      "\033.hbase.test.pb.DummyRequest\032\034.hbase.tes" +
-      "t.pb.DummyResponseB_\n6org.apache.hadoop." +
-      "hbase.coprocessor.protobuf.generatedB\037Du" +
-      "mmyRegionServerEndpointProtos\210\001\001\240\001\001"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_hbase_test_pb_DummyRequest_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_hbase_test_pb_DummyRequest_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_test_pb_DummyRequest_descriptor,
-              new java.lang.String[] { });
-          internal_static_hbase_test_pb_DummyResponse_descriptor =
-            getDescriptor().getMessageTypes().get(1);
-          internal_static_hbase_test_pb_DummyResponse_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_test_pb_DummyResponse_descriptor,
-              new java.lang.String[] { "Value", });
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[28/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMask.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMask.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMask.java
deleted file mode 100644
index df42bca..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMask.java
+++ /dev/null
@@ -1,903 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/field_mask.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * `FieldMask` represents a set of symbolic field paths, for example:
- *     paths: "f.a"
- *     paths: "f.b.d"
- * Here `f` represents a field in some root message, `a` and `b`
- * fields in the message found in `f`, and `d` a field found in the
- * message in `f.b`.
- * Field masks are used to specify a subset of fields that should be
- * returned by a get operation or modified by an update operation.
- * Field masks also have a custom JSON encoding (see below).
- * # Field Masks in Projections
- * When used in the context of a projection, a response message or
- * sub-message is filtered by the API to only contain those fields as
- * specified in the mask. For example, if the mask in the previous
- * example is applied to a response message as follows:
- *     f {
- *       a : 22
- *       b {
- *         d : 1
- *         x : 2
- *       }
- *       y : 13
- *     }
- *     z: 8
- * The result will not contain specific values for fields x,y and z
- * (their value will be set to the default, and omitted in proto text
- * output):
- *     f {
- *       a : 22
- *       b {
- *         d : 1
- *       }
- *     }
- * A repeated field is not allowed except at the last position of a
- * paths string.
- * If a FieldMask object is not present in a get operation, the
- * operation applies to all fields (as if a FieldMask of all fields
- * had been specified).
- * Note that a field mask does not necessarily apply to the
- * top-level response message. In case of a REST get operation, the
- * field mask applies directly to the response, but in case of a REST
- * list operation, the mask instead applies to each individual message
- * in the returned resource list. In case of a REST custom method,
- * other definitions may be used. Where the mask applies will be
- * clearly documented together with its declaration in the API.  In
- * any case, the effect on the returned resource/resources is required
- * behavior for APIs.
- * # Field Masks in Update Operations
- * A field mask in update operations specifies which fields of the
- * targeted resource are going to be updated. The API is required
- * to only change the values of the fields as specified in the mask
- * and leave the others untouched. If a resource is passed in to
- * describe the updated values, the API ignores the values of all
- * fields not covered by the mask.
- * If a repeated field is specified for an update operation, the existing
- * repeated values in the target resource will be overwritten by the new values.
- * Note that a repeated field is only allowed in the last position of a `paths`
- * string.
- * If a sub-message is specified in the last position of the field mask for an
- * update operation, then the existing sub-message in the target resource is
- * overwritten. Given the target message:
- *     f {
- *       b {
- *         d : 1
- *         x : 2
- *       }
- *       c : 1
- *     }
- * And an update message:
- *     f {
- *       b {
- *         d : 10
- *       }
- *     }
- * then if the field mask is:
- *  paths: "f.b"
- * then the result will be:
- *     f {
- *       b {
- *         d : 10
- *       }
- *       c : 1
- *     }
- * However, if the update mask was:
- *  paths: "f.b.d"
- * then the result would be:
- *     f {
- *       b {
- *         d : 10
- *         x : 2
- *       }
- *       c : 1
- *     }
- * In order to reset a field's value to the default, the field must
- * be in the mask and set to the default value in the provided resource.
- * Hence, in order to reset all fields of a resource, provide a default
- * instance of the resource and set all fields in the mask, or do
- * not provide a mask as described below.
- * If a field mask is not present on update, the operation applies to
- * all fields (as if a field mask of all fields has been specified).
- * Note that in the presence of schema evolution, this may mean that
- * fields the client does not know and has therefore not filled into
- * the request will be reset to their default. If this is unwanted
- * behavior, a specific service may require a client to always specify
- * a field mask, producing an error if not.
- * As with get operations, the location of the resource which
- * describes the updated values in the request message depends on the
- * operation kind. In any case, the effect of the field mask is
- * required to be honored by the API.
- * ## Considerations for HTTP REST
- * The HTTP kind of an update operation which uses a field mask must
- * be set to PATCH instead of PUT in order to satisfy HTTP semantics
- * (PUT must only be used for full updates).
- * # JSON Encoding of Field Masks
- * In JSON, a field mask is encoded as a single string where paths are
- * separated by a comma. Fields name in each path are converted
- * to/from lower-camel naming conventions.
- * As an example, consider the following message declarations:
- *     message Profile {
- *       User user = 1;
- *       Photo photo = 2;
- *     }
- *     message User {
- *       string display_name = 1;
- *       string address = 2;
- *     }
- * In proto a field mask for `Profile` may look as such:
- *     mask {
- *       paths: "user.display_name"
- *       paths: "photo"
- *     }
- * In JSON, the same mask is represented as below:
- *     {
- *       mask: "user.displayName,photo"
- *     }
- * # Field Masks and Oneof Fields
- * Field masks treat fields in oneofs just as regular fields. Consider the
- * following message:
- *     message SampleMessage {
- *       oneof test_oneof {
- *         string name = 4;
- *         SubMessage sub_message = 9;
- *       }
- *     }
- * The field mask can be:
- *     mask {
- *       paths: "name"
- *     }
- * Or:
- *     mask {
- *       paths: "sub_message"
- *     }
- * Note that oneof type names ("test_oneof" in this case) cannot be used in
- * paths.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.FieldMask}
- */
-public  final class FieldMask extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.FieldMask)
-    FieldMaskOrBuilder {
-  // Use FieldMask.newBuilder() to construct.
-  private FieldMask(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private FieldMask() {
-    paths_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private FieldMask(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-            if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-              paths_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList();
-              mutable_bitField0_ |= 0x00000001;
-            }
-            paths_.add(s);
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-        paths_ = paths_.getUnmodifiableView();
-      }
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMaskProto.internal_static_google_protobuf_FieldMask_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMaskProto.internal_static_google_protobuf_FieldMask_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask.Builder.class);
-  }
-
-  public static final int PATHS_FIELD_NUMBER = 1;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList paths_;
-  /**
-   * <pre>
-   * The set of field mask paths.
-   * </pre>
-   *
-   * <code>repeated string paths = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
-      getPathsList() {
-    return paths_;
-  }
-  /**
-   * <pre>
-   * The set of field mask paths.
-   * </pre>
-   *
-   * <code>repeated string paths = 1;</code>
-   */
-  public int getPathsCount() {
-    return paths_.size();
-  }
-  /**
-   * <pre>
-   * The set of field mask paths.
-   * </pre>
-   *
-   * <code>repeated string paths = 1;</code>
-   */
-  public java.lang.String getPaths(int index) {
-    return paths_.get(index);
-  }
-  /**
-   * <pre>
-   * The set of field mask paths.
-   * </pre>
-   *
-   * <code>repeated string paths = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getPathsBytes(int index) {
-    return paths_.getByteString(index);
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    for (int i = 0; i < paths_.size(); i++) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, paths_.getRaw(i));
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    {
-      int dataSize = 0;
-      for (int i = 0; i < paths_.size(); i++) {
-        dataSize += computeStringSizeNoTag(paths_.getRaw(i));
-      }
-      size += dataSize;
-      size += 1 * getPathsList().size();
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask) obj;
-
-    boolean result = true;
-    result = result && getPathsList()
-        .equals(other.getPathsList());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    if (getPathsCount() > 0) {
-      hash = (37 * hash) + PATHS_FIELD_NUMBER;
-      hash = (53 * hash) + getPathsList().hashCode();
-    }
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * `FieldMask` represents a set of symbolic field paths, for example:
-   *     paths: "f.a"
-   *     paths: "f.b.d"
-   * Here `f` represents a field in some root message, `a` and `b`
-   * fields in the message found in `f`, and `d` a field found in the
-   * message in `f.b`.
-   * Field masks are used to specify a subset of fields that should be
-   * returned by a get operation or modified by an update operation.
-   * Field masks also have a custom JSON encoding (see below).
-   * # Field Masks in Projections
-   * When used in the context of a projection, a response message or
-   * sub-message is filtered by the API to only contain those fields as
-   * specified in the mask. For example, if the mask in the previous
-   * example is applied to a response message as follows:
-   *     f {
-   *       a : 22
-   *       b {
-   *         d : 1
-   *         x : 2
-   *       }
-   *       y : 13
-   *     }
-   *     z: 8
-   * The result will not contain specific values for fields x,y and z
-   * (their value will be set to the default, and omitted in proto text
-   * output):
-   *     f {
-   *       a : 22
-   *       b {
-   *         d : 1
-   *       }
-   *     }
-   * A repeated field is not allowed except at the last position of a
-   * paths string.
-   * If a FieldMask object is not present in a get operation, the
-   * operation applies to all fields (as if a FieldMask of all fields
-   * had been specified).
-   * Note that a field mask does not necessarily apply to the
-   * top-level response message. In case of a REST get operation, the
-   * field mask applies directly to the response, but in case of a REST
-   * list operation, the mask instead applies to each individual message
-   * in the returned resource list. In case of a REST custom method,
-   * other definitions may be used. Where the mask applies will be
-   * clearly documented together with its declaration in the API.  In
-   * any case, the effect on the returned resource/resources is required
-   * behavior for APIs.
-   * # Field Masks in Update Operations
-   * A field mask in update operations specifies which fields of the
-   * targeted resource are going to be updated. The API is required
-   * to only change the values of the fields as specified in the mask
-   * and leave the others untouched. If a resource is passed in to
-   * describe the updated values, the API ignores the values of all
-   * fields not covered by the mask.
-   * If a repeated field is specified for an update operation, the existing
-   * repeated values in the target resource will be overwritten by the new values.
-   * Note that a repeated field is only allowed in the last position of a `paths`
-   * string.
-   * If a sub-message is specified in the last position of the field mask for an
-   * update operation, then the existing sub-message in the target resource is
-   * overwritten. Given the target message:
-   *     f {
-   *       b {
-   *         d : 1
-   *         x : 2
-   *       }
-   *       c : 1
-   *     }
-   * And an update message:
-   *     f {
-   *       b {
-   *         d : 10
-   *       }
-   *     }
-   * then if the field mask is:
-   *  paths: "f.b"
-   * then the result will be:
-   *     f {
-   *       b {
-   *         d : 10
-   *       }
-   *       c : 1
-   *     }
-   * However, if the update mask was:
-   *  paths: "f.b.d"
-   * then the result would be:
-   *     f {
-   *       b {
-   *         d : 10
-   *         x : 2
-   *       }
-   *       c : 1
-   *     }
-   * In order to reset a field's value to the default, the field must
-   * be in the mask and set to the default value in the provided resource.
-   * Hence, in order to reset all fields of a resource, provide a default
-   * instance of the resource and set all fields in the mask, or do
-   * not provide a mask as described below.
-   * If a field mask is not present on update, the operation applies to
-   * all fields (as if a field mask of all fields has been specified).
-   * Note that in the presence of schema evolution, this may mean that
-   * fields the client does not know and has therefore not filled into
-   * the request will be reset to their default. If this is unwanted
-   * behavior, a specific service may require a client to always specify
-   * a field mask, producing an error if not.
-   * As with get operations, the location of the resource which
-   * describes the updated values in the request message depends on the
-   * operation kind. In any case, the effect of the field mask is
-   * required to be honored by the API.
-   * ## Considerations for HTTP REST
-   * The HTTP kind of an update operation which uses a field mask must
-   * be set to PATCH instead of PUT in order to satisfy HTTP semantics
-   * (PUT must only be used for full updates).
-   * # JSON Encoding of Field Masks
-   * In JSON, a field mask is encoded as a single string where paths are
-   * separated by a comma. Fields name in each path are converted
-   * to/from lower-camel naming conventions.
-   * As an example, consider the following message declarations:
-   *     message Profile {
-   *       User user = 1;
-   *       Photo photo = 2;
-   *     }
-   *     message User {
-   *       string display_name = 1;
-   *       string address = 2;
-   *     }
-   * In proto a field mask for `Profile` may look as such:
-   *     mask {
-   *       paths: "user.display_name"
-   *       paths: "photo"
-   *     }
-   * In JSON, the same mask is represented as below:
-   *     {
-   *       mask: "user.displayName,photo"
-   *     }
-   * # Field Masks and Oneof Fields
-   * Field masks treat fields in oneofs just as regular fields. Consider the
-   * following message:
-   *     message SampleMessage {
-   *       oneof test_oneof {
-   *         string name = 4;
-   *         SubMessage sub_message = 9;
-   *       }
-   *     }
-   * The field mask can be:
-   *     mask {
-   *       paths: "name"
-   *     }
-   * Or:
-   *     mask {
-   *       paths: "sub_message"
-   *     }
-   * Note that oneof type names ("test_oneof" in this case) cannot be used in
-   * paths.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.FieldMask}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.FieldMask)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMaskOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMaskProto.internal_static_google_protobuf_FieldMask_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMaskProto.internal_static_google_protobuf_FieldMask_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      paths_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-      bitField0_ = (bitField0_ & ~0x00000001);
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMaskProto.internal_static_google_protobuf_FieldMask_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask(this);
-      int from_bitField0_ = bitField0_;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        paths_ = paths_.getUnmodifiableView();
-        bitField0_ = (bitField0_ & ~0x00000001);
-      }
-      result.paths_ = paths_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask.getDefaultInstance()) return this;
-      if (!other.paths_.isEmpty()) {
-        if (paths_.isEmpty()) {
-          paths_ = other.paths_;
-          bitField0_ = (bitField0_ & ~0x00000001);
-        } else {
-          ensurePathsIsMutable();
-          paths_.addAll(other.paths_);
-        }
-        onChanged();
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringList paths_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-    private void ensurePathsIsMutable() {
-      if (!((bitField0_ & 0x00000001) == 0x00000001)) {
-        paths_ = new org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList(paths_);
-        bitField0_ |= 0x00000001;
-       }
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolStringList
-        getPathsList() {
-      return paths_.getUnmodifiableView();
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public int getPathsCount() {
-      return paths_.size();
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public java.lang.String getPaths(int index) {
-      return paths_.get(index);
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getPathsBytes(int index) {
-      return paths_.getByteString(index);
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public Builder setPaths(
-        int index, java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  ensurePathsIsMutable();
-      paths_.set(index, value);
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public Builder addPaths(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  ensurePathsIsMutable();
-      paths_.add(value);
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public Builder addAllPaths(
-        java.lang.Iterable<java.lang.String> values) {
-      ensurePathsIsMutable();
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.addAll(
-          values, paths_);
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public Builder clearPaths() {
-      paths_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyStringArrayList.EMPTY;
-      bitField0_ = (bitField0_ & ~0x00000001);
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The set of field mask paths.
-     * </pre>
-     *
-     * <code>repeated string paths = 1;</code>
-     */
-    public Builder addPathsBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      ensurePathsIsMutable();
-      paths_.add(value);
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.FieldMask)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.FieldMask)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FieldMask>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<FieldMask>() {
-    public FieldMask parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new FieldMask(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FieldMask> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<FieldMask> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.FieldMask getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMaskOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMaskOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMaskOrBuilder.java
deleted file mode 100644
index da0e875..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMaskOrBuilder.java
+++ /dev/null
@@ -1,44 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/field_mask.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface FieldMaskOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.FieldMask)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The set of field mask paths.
-   * </pre>
-   *
-   * <code>repeated string paths = 1;</code>
-   */
-  java.util.List<java.lang.String>
-      getPathsList();
-  /**
-   * <pre>
-   * The set of field mask paths.
-   * </pre>
-   *
-   * <code>repeated string paths = 1;</code>
-   */
-  int getPathsCount();
-  /**
-   * <pre>
-   * The set of field mask paths.
-   * </pre>
-   *
-   * <code>repeated string paths = 1;</code>
-   */
-  java.lang.String getPaths(int index);
-  /**
-   * <pre>
-   * The set of field mask paths.
-   * </pre>
-   *
-   * <code>repeated string paths = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getPathsBytes(int index);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMaskProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMaskProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMaskProto.java
deleted file mode 100644
index 4e978bf..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldMaskProto.java
+++ /dev/null
@@ -1,59 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/field_mask.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class FieldMaskProto {
-  private FieldMaskProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_FieldMask_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_FieldMask_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n google/protobuf/field_mask.proto\022\017goog" +
-      "le.protobuf\"\032\n\tFieldMask\022\r\n\005paths\030\001 \003(\tB" +
-      "\211\001\n\023com.google.protobufB\016FieldMaskProtoP" +
-      "\001Z9google.golang.org/genproto/protobuf/f" +
-      "ield_mask;field_mask\242\002\003GPB\252\002\036Google.Prot" +
-      "obuf.WellKnownTypesb\006proto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_google_protobuf_FieldMask_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_FieldMask_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_FieldMask_descriptor,
-        new java.lang.String[] { "Paths", });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldOrBuilder.java
deleted file mode 100644
index 1bd3ada..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldOrBuilder.java
+++ /dev/null
@@ -1,189 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface FieldOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Field)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The field type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Field.Kind kind = 1;</code>
-   */
-  int getKindValue();
-  /**
-   * <pre>
-   * The field type.
-   * </pre>
-   *
-   * <code>.google.protobuf.Field.Kind kind = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Kind getKind();
-
-  /**
-   * <pre>
-   * The field cardinality.
-   * </pre>
-   *
-   * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-   */
-  int getCardinalityValue();
-  /**
-   * <pre>
-   * The field cardinality.
-   * </pre>
-   *
-   * <code>.google.protobuf.Field.Cardinality cardinality = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Field.Cardinality getCardinality();
-
-  /**
-   * <pre>
-   * The field number.
-   * </pre>
-   *
-   * <code>int32 number = 3;</code>
-   */
-  int getNumber();
-
-  /**
-   * <pre>
-   * The field name.
-   * </pre>
-   *
-   * <code>string name = 4;</code>
-   */
-  java.lang.String getName();
-  /**
-   * <pre>
-   * The field name.
-   * </pre>
-   *
-   * <code>string name = 4;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getNameBytes();
-
-  /**
-   * <pre>
-   * The field type URL, without the scheme, for message or enumeration
-   * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-   * </pre>
-   *
-   * <code>string type_url = 6;</code>
-   */
-  java.lang.String getTypeUrl();
-  /**
-   * <pre>
-   * The field type URL, without the scheme, for message or enumeration
-   * types. Example: `"type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp"`.
-   * </pre>
-   *
-   * <code>string type_url = 6;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getTypeUrlBytes();
-
-  /**
-   * <pre>
-   * The index of the field type in `Type.oneofs`, for message or enumeration
-   * types. The first type has index 1; zero means the type is not in the list.
-   * </pre>
-   *
-   * <code>int32 oneof_index = 7;</code>
-   */
-  int getOneofIndex();
-
-  /**
-   * <pre>
-   * Whether to use alternative packed wire representation.
-   * </pre>
-   *
-   * <code>bool packed = 8;</code>
-   */
-  boolean getPacked();
-
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> 
-      getOptionsList();
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index);
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  int getOptionsCount();
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> 
-      getOptionsOrBuilderList();
-  /**
-   * <pre>
-   * The protocol buffer options.
-   * </pre>
-   *
-   * <code>repeated .google.protobuf.Option options = 9;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder(
-      int index);
-
-  /**
-   * <pre>
-   * The field JSON name.
-   * </pre>
-   *
-   * <code>string json_name = 10;</code>
-   */
-  java.lang.String getJsonName();
-  /**
-   * <pre>
-   * The field JSON name.
-   * </pre>
-   *
-   * <code>string json_name = 10;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getJsonNameBytes();
-
-  /**
-   * <pre>
-   * The string value of the default value of this field. Proto2 syntax only.
-   * </pre>
-   *
-   * <code>string default_value = 11;</code>
-   */
-  java.lang.String getDefaultValue();
-  /**
-   * <pre>
-   * The string value of the default value of this field. Proto2 syntax only.
-   * </pre>
-   *
-   * <code>string default_value = 11;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getDefaultValueBytes();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldSet.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldSet.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldSet.java
deleted file mode 100644
index a1bf553..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FieldSet.java
+++ /dev/null
@@ -1,909 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.LazyField.LazyIterator;
-
-import java.io.IOException;
-import java.util.ArrayList;
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-
-/**
- * A class which represents an arbitrary set of fields of some message type.
- * This is used to implement {@link DynamicMessage}, and also to represent
- * extensions in {@link GeneratedMessage}.  This class is package-private,
- * since outside users should probably be using {@link DynamicMessage}.
- *
- * @author kenton@google.com Kenton Varda
- */
-final class FieldSet<FieldDescriptorType extends
-      FieldSet.FieldDescriptorLite<FieldDescriptorType>> {
-  /**
-   * Interface for a FieldDescriptor or lite extension descriptor.  This
-   * prevents FieldSet from depending on {@link Descriptors.FieldDescriptor}.
-   */
-  public interface FieldDescriptorLite<T extends FieldDescriptorLite<T>>
-      extends Comparable<T> {
-    int getNumber();
-    WireFormat.FieldType getLiteType();
-    WireFormat.JavaType getLiteJavaType();
-    boolean isRepeated();
-    boolean isPacked();
-    Internal.EnumLiteMap<?> getEnumType();
-
-    // If getLiteJavaType() == MESSAGE, this merges a message object of the
-    // type into a builder of the type.  Returns {@code to}.
-    MessageLite.Builder internalMergeFrom(
-        MessageLite.Builder to, MessageLite from);
-  }
-
-  private final SmallSortedMap<FieldDescriptorType, Object> fields;
-  private boolean isImmutable;
-  private boolean hasLazyField = false;
-
-  /** Construct a new FieldSet. */
-  private FieldSet() {
-    this.fields = SmallSortedMap.newFieldMap(16);
-  }
-
-  /**
-   * Construct an empty FieldSet.  This is only used to initialize
-   * DEFAULT_INSTANCE.
-   */
-  private FieldSet(final boolean dummy) {
-    this.fields = SmallSortedMap.newFieldMap(0);
-    makeImmutable();
-  }
-
-  /** Construct a new FieldSet. */
-  public static <T extends FieldSet.FieldDescriptorLite<T>>
-      FieldSet<T> newFieldSet() {
-    return new FieldSet<T>();
-  }
-
-  /** Get an immutable empty FieldSet. */
-  @SuppressWarnings("unchecked")
-  public static <T extends FieldSet.FieldDescriptorLite<T>>
-      FieldSet<T> emptySet() {
-    return DEFAULT_INSTANCE;
-  }
-  @SuppressWarnings("rawtypes")
-  private static final FieldSet DEFAULT_INSTANCE = new FieldSet(true);
-
-  /** Make this FieldSet immutable from this point forward. */
-  @SuppressWarnings("unchecked")
-  public void makeImmutable() {
-    if (isImmutable) {
-      return;
-    }
-    fields.makeImmutable();
-    isImmutable = true;
-  }
-
-  /**
-   * Returns whether the FieldSet is immutable. This is true if it is the
-   * {@link #emptySet} or if {@link #makeImmutable} were called.
-   *
-   * @return whether the FieldSet is immutable.
-   */
-  public boolean isImmutable() {
-    return isImmutable;
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (this == o) {
-      return true;
-    }
-
-    if (!(o instanceof FieldSet)) {
-      return false;
-    }
-
-    FieldSet<?> other = (FieldSet<?>) o;
-    return fields.equals(other.fields);
-  }
-
-  @Override
-  public int hashCode() {
-    return fields.hashCode();
-  }
-
-  /**
-   * Clones the FieldSet. The returned FieldSet will be mutable even if the
-   * original FieldSet was immutable.
-   *
-   * @return the newly cloned FieldSet
-   */
-  @Override
-  public FieldSet<FieldDescriptorType> clone() {
-    // We can't just call fields.clone because List objects in the map
-    // should not be shared.
-    FieldSet<FieldDescriptorType> clone = FieldSet.newFieldSet();
-    for (int i = 0; i < fields.getNumArrayEntries(); i++) {
-      Map.Entry<FieldDescriptorType, Object> entry = fields.getArrayEntryAt(i);
-      FieldDescriptorType descriptor = entry.getKey();
-      clone.setField(descriptor, entry.getValue());
-    }
-    for (Map.Entry<FieldDescriptorType, Object> entry :
-             fields.getOverflowEntries()) {
-      FieldDescriptorType descriptor = entry.getKey();
-      clone.setField(descriptor, entry.getValue());
-    }
-    clone.hasLazyField = hasLazyField;
-    return clone;
-  }
-
-
-  // =================================================================
-
-  /** See {@link Message.Builder#clear()}. */
-  public void clear() {
-    fields.clear();
-    hasLazyField = false;
-  }
-
-  /**
-   * Get a simple map containing all the fields.
-   */
-  public Map<FieldDescriptorType, Object> getAllFields() {
-    if (hasLazyField) {
-      SmallSortedMap<FieldDescriptorType, Object> result =
-          SmallSortedMap.newFieldMap(16);
-      for (int i = 0; i < fields.getNumArrayEntries(); i++) {
-        cloneFieldEntry(result, fields.getArrayEntryAt(i));
-      }
-      for (Map.Entry<FieldDescriptorType, Object> entry :
-          fields.getOverflowEntries()) {
-        cloneFieldEntry(result, entry);
-      }
-      if (fields.isImmutable()) {
-        result.makeImmutable();
-      }
-      return result;
-    }
-    return fields.isImmutable() ? fields : Collections.unmodifiableMap(fields);
-  }
-
-  private void cloneFieldEntry(Map<FieldDescriptorType, Object> map,
-      Map.Entry<FieldDescriptorType, Object> entry) {
-    FieldDescriptorType key = entry.getKey();
-    Object value = entry.getValue();
-    if (value instanceof LazyField) {
-      map.put(key, ((LazyField) value).getValue());
-    } else {
-      map.put(key, value);
-    }
-  }
-
-  /**
-   * Get an iterator to the field map. This iterator should not be leaked out
-   * of the protobuf library as it is not protected from mutation when fields
-   * is not immutable.
-   */
-  public Iterator<Map.Entry<FieldDescriptorType, Object>> iterator() {
-    if (hasLazyField) {
-      return new LazyIterator<FieldDescriptorType>(
-          fields.entrySet().iterator());
-    }
-    return fields.entrySet().iterator();
-  }
-
-
-  /**
-   * Useful for implementing
-   * {@link Message#hasField(Descriptors.FieldDescriptor)}.
-   */
-  public boolean hasField(final FieldDescriptorType descriptor) {
-    if (descriptor.isRepeated()) {
-      throw new IllegalArgumentException(
-        "hasField() can only be called on non-repeated fields.");
-    }
-
-    return fields.get(descriptor) != null;
-  }
-
-  /**
-   * Useful for implementing
-   * {@link Message#getField(Descriptors.FieldDescriptor)}.  This method
-   * returns {@code null} if the field is not set; in this case it is up
-   * to the caller to fetch the field's default value.
-   */
-  public Object getField(final FieldDescriptorType descriptor) {
-    Object o = fields.get(descriptor);
-    if (o instanceof LazyField) {
-      return ((LazyField) o).getValue();
-    }
-    return o;
-  }
-
-  /**
-   * Useful for implementing
-   * {@link Message.Builder#setField(Descriptors.FieldDescriptor,Object)}.
-   */
-  @SuppressWarnings({"unchecked", "rawtypes"})
-  public void setField(final FieldDescriptorType descriptor,
-                       Object value) {
-    if (descriptor.isRepeated()) {
-      if (!(value instanceof List)) {
-        throw new IllegalArgumentException(
-          "Wrong object type used with protocol message reflection.");
-      }
-
-      // Wrap the contents in a new list so that the caller cannot change
-      // the list's contents after setting it.
-      final List newList = new ArrayList();
-      newList.addAll((List) value);
-      for (final Object element : newList) {
-        verifyType(descriptor.getLiteType(), element);
-      }
-      value = newList;
-    } else {
-      verifyType(descriptor.getLiteType(), value);
-    }
-
-    if (value instanceof LazyField) {
-      hasLazyField = true;
-    }
-    fields.put(descriptor, value);
-  }
-
-  /**
-   * Useful for implementing
-   * {@link Message.Builder#clearField(Descriptors.FieldDescriptor)}.
-   */
-  public void clearField(final FieldDescriptorType descriptor) {
-    fields.remove(descriptor);
-    if (fields.isEmpty()) {
-      hasLazyField = false;
-    }
-  }
-
-  /**
-   * Useful for implementing
-   * {@link Message#getRepeatedFieldCount(Descriptors.FieldDescriptor)}.
-   */
-  public int getRepeatedFieldCount(final FieldDescriptorType descriptor) {
-    if (!descriptor.isRepeated()) {
-      throw new IllegalArgumentException(
-        "getRepeatedField() can only be called on repeated fields.");
-    }
-
-    final Object value = getField(descriptor);
-    if (value == null) {
-      return 0;
-    } else {
-      return ((List<?>) value).size();
-    }
-  }
-
-  /**
-   * Useful for implementing
-   * {@link Message#getRepeatedField(Descriptors.FieldDescriptor,int)}.
-   */
-  public Object getRepeatedField(final FieldDescriptorType descriptor,
-                                 final int index) {
-    if (!descriptor.isRepeated()) {
-      throw new IllegalArgumentException(
-        "getRepeatedField() can only be called on repeated fields.");
-    }
-
-    final Object value = getField(descriptor);
-
-    if (value == null) {
-      throw new IndexOutOfBoundsException();
-    } else {
-      return ((List<?>) value).get(index);
-    }
-  }
-
-  /**
-   * Useful for implementing
-   * {@link Message.Builder#setRepeatedField(Descriptors.FieldDescriptor,int,Object)}.
-   */
-  @SuppressWarnings("unchecked")
-  public void setRepeatedField(final FieldDescriptorType descriptor,
-                               final int index,
-                               final Object value) {
-    if (!descriptor.isRepeated()) {
-      throw new IllegalArgumentException(
-        "getRepeatedField() can only be called on repeated fields.");
-    }
-
-    final Object list = getField(descriptor);
-    if (list == null) {
-      throw new IndexOutOfBoundsException();
-    }
-
-    verifyType(descriptor.getLiteType(), value);
-    ((List<Object>) list).set(index, value);
-  }
-
-  /**
-   * Useful for implementing
-   * {@link Message.Builder#addRepeatedField(Descriptors.FieldDescriptor,Object)}.
-   */
-  @SuppressWarnings("unchecked")
-  public void addRepeatedField(final FieldDescriptorType descriptor,
-                               final Object value) {
-    if (!descriptor.isRepeated()) {
-      throw new IllegalArgumentException(
-        "addRepeatedField() can only be called on repeated fields.");
-    }
-
-    verifyType(descriptor.getLiteType(), value);
-
-    final Object existingValue = getField(descriptor);
-    List<Object> list;
-    if (existingValue == null) {
-      list = new ArrayList<Object>();
-      fields.put(descriptor, list);
-    } else {
-      list = (List<Object>) existingValue;
-    }
-
-    list.add(value);
-  }
-
-  /**
-   * Verifies that the given object is of the correct type to be a valid
-   * value for the given field.  (For repeated fields, this checks if the
-   * object is the right type to be one element of the field.)
-   *
-   * @throws IllegalArgumentException The value is not of the right type.
-   */
-  private static void verifyType(final WireFormat.FieldType type,
-                                 final Object value) {
-    if (value == null) {
-      throw new NullPointerException();
-    }
-
-    boolean isValid = false;
-    switch (type.getJavaType()) {
-      case INT:          isValid = value instanceof Integer   ; break;
-      case LONG:         isValid = value instanceof Long      ; break;
-      case FLOAT:        isValid = value instanceof Float     ; break;
-      case DOUBLE:       isValid = value instanceof Double    ; break;
-      case BOOLEAN:      isValid = value instanceof Boolean   ; break;
-      case STRING:       isValid = value instanceof String    ; break;
-      case BYTE_STRING:
-        isValid = value instanceof ByteString || value instanceof byte[];
-        break;
-      case ENUM:
-        // TODO(kenton):  Caller must do type checking here, I guess.
-        isValid =
-            (value instanceof Integer || value instanceof Internal.EnumLite);
-        break;
-      case MESSAGE:
-        // TODO(kenton):  Caller must do type checking here, I guess.
-        isValid =
-            (value instanceof MessageLite) || (value instanceof LazyField);
-        break;
-    }
-
-    if (!isValid) {
-      // TODO(kenton):  When chaining calls to setField(), it can be hard to
-      //   tell from the stack trace which exact call failed, since the whole
-      //   chain is considered one line of code.  It would be nice to print
-      //   more information here, e.g. naming the field.  We used to do that.
-      //   But we can't now that FieldSet doesn't use descriptors.  Maybe this
-      //   isn't a big deal, though, since it would only really apply when using
-      //   reflection and generally people don't chain reflection setters.
-      throw new IllegalArgumentException(
-        "Wrong object type used with protocol message reflection.");
-    }
-  }
-
-  // =================================================================
-  // Parsing and serialization
-
-  /**
-   * See {@link Message#isInitialized()}.  Note:  Since {@code FieldSet}
-   * itself does not have any way of knowing about required fields that
-   * aren't actually present in the set, it is up to the caller to check
-   * that all required fields are present.
-   */
-  public boolean isInitialized() {
-    for (int i = 0; i < fields.getNumArrayEntries(); i++) {
-      if (!isInitialized(fields.getArrayEntryAt(i))) {
-        return false;
-      }
-    }
-    for (final Map.Entry<FieldDescriptorType, Object> entry :
-             fields.getOverflowEntries()) {
-      if (!isInitialized(entry)) {
-        return false;
-      }
-    }
-    return true;
-  }
-
-  @SuppressWarnings("unchecked")
-  private boolean isInitialized(
-      final Map.Entry<FieldDescriptorType, Object> entry) {
-    final FieldDescriptorType descriptor = entry.getKey();
-    if (descriptor.getLiteJavaType() == WireFormat.JavaType.MESSAGE) {
-      if (descriptor.isRepeated()) {
-        for (final MessageLite element:
-                 (List<MessageLite>) entry.getValue()) {
-          if (!element.isInitialized()) {
-            return false;
-          }
-        }
-      } else {
-        Object value = entry.getValue();
-        if (value instanceof MessageLite) {
-          if (!((MessageLite) value).isInitialized()) {
-            return false;
-          }
-        } else if (value instanceof LazyField) {
-          return true;
-        } else {
-          throw new IllegalArgumentException(
-              "Wrong object type used with protocol message reflection.");
-        }
-      }
-    }
-    return true;
-  }
-
-  /**
-   * Given a field type, return the wire type.
-   *
-   * @returns One of the {@code WIRETYPE_} constants defined in
-   *          {@link WireFormat}.
-   */
-  static int getWireFormatForFieldType(final WireFormat.FieldType type,
-                                       boolean isPacked) {
-    if (isPacked) {
-      return WireFormat.WIRETYPE_LENGTH_DELIMITED;
-    } else {
-      return type.getWireType();
-    }
-  }
-
-  /**
-   * Like {@link Message.Builder#mergeFrom(Message)}, but merges from another
-   * {@link FieldSet}.
-   */
-  public void mergeFrom(final FieldSet<FieldDescriptorType> other) {
-    for (int i = 0; i < other.fields.getNumArrayEntries(); i++) {
-      mergeFromField(other.fields.getArrayEntryAt(i));
-    }
-    for (final Map.Entry<FieldDescriptorType, Object> entry :
-             other.fields.getOverflowEntries()) {
-      mergeFromField(entry);
-    }
-  }
-
-  private Object cloneIfMutable(Object value) {
-    if (value instanceof byte[]) {
-      byte[] bytes = (byte[]) value;
-      byte[] copy = new byte[bytes.length];
-      System.arraycopy(bytes, 0, copy, 0, bytes.length);
-      return copy;
-    } else {
-      return value;
-    }
-  }
-
-  @SuppressWarnings({"unchecked", "rawtypes"})
-  private void mergeFromField(
-      final Map.Entry<FieldDescriptorType, Object> entry) {
-    final FieldDescriptorType descriptor = entry.getKey();
-    Object otherValue = entry.getValue();
-    if (otherValue instanceof LazyField) {
-      otherValue = ((LazyField) otherValue).getValue();
-    }
-
-    if (descriptor.isRepeated()) {
-      Object value = getField(descriptor);
-      if (value == null) {
-        value = new ArrayList();
-      }
-      for (Object element : (List) otherValue) {
-        ((List) value).add(cloneIfMutable(element));
-      }
-      fields.put(descriptor, value);
-    } else if (descriptor.getLiteJavaType() == WireFormat.JavaType.MESSAGE) {
-      Object value = getField(descriptor);
-      if (value == null) {
-        fields.put(descriptor, cloneIfMutable(otherValue));
-      } else {
-        // Merge the messages.
-          value = descriptor.internalMergeFrom(
-                ((MessageLite) value).toBuilder(), (MessageLite) otherValue)
-                .build();
-
-        fields.put(descriptor, value);
-      }
-    } else {
-      fields.put(descriptor, cloneIfMutable(otherValue));
-    }
-  }
-
-  // TODO(kenton):  Move static parsing and serialization methods into some
-  //   other class.  Probably WireFormat.
-
-  /**
-   * Read a field of any primitive type for immutable messages from a
-   * CodedInputStream. Enums, groups, and embedded messages are not handled by
-   * this method.
-   *
-   * @param input The stream from which to read.
-   * @param type Declared type of the field.
-   * @param checkUtf8 When true, check that the input is valid utf8.
-   * @return An object representing the field's value, of the exact
-   *         type which would be returned by
-   *         {@link Message#getField(Descriptors.FieldDescriptor)} for
-   *         this field.
-   */
-  public static Object readPrimitiveField(
-      CodedInputStream input,
-      final WireFormat.FieldType type,
-      boolean checkUtf8) throws IOException {
-    if (checkUtf8) {
-      return WireFormat.readPrimitiveField(input, type,
-          WireFormat.Utf8Validation.STRICT);
-    } else {
-      return WireFormat.readPrimitiveField(input, type,
-          WireFormat.Utf8Validation.LOOSE);
-    }
-  }
-
-
-  /** See {@link Message#writeTo(CodedOutputStream)}. */
-  public void writeTo(final CodedOutputStream output)
-                      throws IOException {
-    for (int i = 0; i < fields.getNumArrayEntries(); i++) {
-      final Map.Entry<FieldDescriptorType, Object> entry =
-          fields.getArrayEntryAt(i);
-      writeField(entry.getKey(), entry.getValue(), output);
-    }
-    for (final Map.Entry<FieldDescriptorType, Object> entry :
-         fields.getOverflowEntries()) {
-      writeField(entry.getKey(), entry.getValue(), output);
-    }
-  }
-
-  /**
-   * Like {@link #writeTo} but uses MessageSet wire format.
-   */
-  public void writeMessageSetTo(final CodedOutputStream output)
-                                throws IOException {
-    for (int i = 0; i < fields.getNumArrayEntries(); i++) {
-      writeMessageSetTo(fields.getArrayEntryAt(i), output);
-    }
-    for (final Map.Entry<FieldDescriptorType, Object> entry :
-             fields.getOverflowEntries()) {
-      writeMessageSetTo(entry, output);
-    }
-  }
-
-  private void writeMessageSetTo(
-      final Map.Entry<FieldDescriptorType, Object> entry,
-      final CodedOutputStream output) throws IOException {
-    final FieldDescriptorType descriptor = entry.getKey();
-    if (descriptor.getLiteJavaType() == WireFormat.JavaType.MESSAGE &&
-        !descriptor.isRepeated() && !descriptor.isPacked()) {
-      Object value = entry.getValue();
-      if (value instanceof LazyField) {
-        value = ((LazyField) value).getValue();
-      }
-      output.writeMessageSetExtension(entry.getKey().getNumber(),
-                                      (MessageLite) value);
-    } else {
-      writeField(descriptor, entry.getValue(), output);
-    }
-  }
-
-  /**
-   * Write a single tag-value pair to the stream.
-   *
-   * @param output The output stream.
-   * @param type   The field's type.
-   * @param number The field's number.
-   * @param value  Object representing the field's value.  Must be of the exact
-   *               type which would be returned by
-   *               {@link Message#getField(Descriptors.FieldDescriptor)} for
-   *               this field.
-   */
-  static void writeElement(
-      final CodedOutputStream output,
-      final WireFormat.FieldType type,
-      final int number,
-      final Object value) throws IOException {
-    // Special case for groups, which need a start and end tag; other fields
-    // can just use writeTag() and writeFieldNoTag().
-    if (type == WireFormat.FieldType.GROUP) {
-        output.writeGroup(number, (MessageLite) value);
-    } else {
-      output.writeTag(number, getWireFormatForFieldType(type, false));
-      writeElementNoTag(output, type, value);
-    }
-  }
-
-  /**
-   * Write a field of arbitrary type, without its tag, to the stream.
-   *
-   * @param output The output stream.
-   * @param type The field's type.
-   * @param value  Object representing the field's value.  Must be of the exact
-   *               type which would be returned by
-   *               {@link Message#getField(Descriptors.FieldDescriptor)} for
-   *               this field.
-   */
-  static void writeElementNoTag(
-      final CodedOutputStream output,
-      final WireFormat.FieldType type,
-      final Object value) throws IOException {
-    switch (type) {
-      case DOUBLE  : output.writeDoubleNoTag  ((Double     ) value); break;
-      case FLOAT   : output.writeFloatNoTag   ((Float      ) value); break;
-      case INT64   : output.writeInt64NoTag   ((Long       ) value); break;
-      case UINT64  : output.writeUInt64NoTag  ((Long       ) value); break;
-      case INT32   : output.writeInt32NoTag   ((Integer    ) value); break;
-      case FIXED64 : output.writeFixed64NoTag ((Long       ) value); break;
-      case FIXED32 : output.writeFixed32NoTag ((Integer    ) value); break;
-      case BOOL    : output.writeBoolNoTag    ((Boolean    ) value); break;
-      case GROUP   : output.writeGroupNoTag   ((MessageLite) value); break;
-      case MESSAGE : output.writeMessageNoTag ((MessageLite) value); break;
-      case STRING:
-        if (value instanceof ByteString) {
-          output.writeBytesNoTag((ByteString) value);
-        } else {
-          output.writeStringNoTag((String) value);
-        }
-        break;
-      case BYTES:
-        if (value instanceof ByteString) {
-          output.writeBytesNoTag((ByteString) value);
-        } else {
-          output.writeByteArrayNoTag((byte[]) value);
-        }
-        break;
-      case UINT32  : output.writeUInt32NoTag  ((Integer    ) value); break;
-      case SFIXED32: output.writeSFixed32NoTag((Integer    ) value); break;
-      case SFIXED64: output.writeSFixed64NoTag((Long       ) value); break;
-      case SINT32  : output.writeSInt32NoTag  ((Integer    ) value); break;
-      case SINT64  : output.writeSInt64NoTag  ((Long       ) value); break;
-
-      case ENUM:
-        if (value instanceof Internal.EnumLite) {
-          output.writeEnumNoTag(((Internal.EnumLite) value).getNumber());
-        } else {
-          output.writeEnumNoTag(((Integer) value).intValue());
-        }
-        break;
-    }
-  }
-
-  /** Write a single field. */
-  public static void writeField(final FieldDescriptorLite<?> descriptor,
-                                final Object value,
-                                final CodedOutputStream output)
-                                throws IOException {
-    WireFormat.FieldType type = descriptor.getLiteType();
-    int number = descriptor.getNumber();
-    if (descriptor.isRepeated()) {
-      final List<?> valueList = (List<?>)value;
-      if (descriptor.isPacked()) {
-        output.writeTag(number, WireFormat.WIRETYPE_LENGTH_DELIMITED);
-        // Compute the total data size so the length can be written.
-        int dataSize = 0;
-        for (final Object element : valueList) {
-          dataSize += computeElementSizeNoTag(type, element);
-        }
-        output.writeRawVarint32(dataSize);
-        // Write the data itself, without any tags.
-        for (final Object element : valueList) {
-          writeElementNoTag(output, type, element);
-        }
-      } else {
-        for (final Object element : valueList) {
-          writeElement(output, type, number, element);
-        }
-      }
-    } else {
-      if (value instanceof LazyField) {
-        writeElement(output, type, number, ((LazyField) value).getValue());
-      } else {
-        writeElement(output, type, number, value);
-      }
-    }
-  }
-
-  /**
-   * See {@link Message#getSerializedSize()}.  It's up to the caller to cache
-   * the resulting size if desired.
-   */
-  public int getSerializedSize() {
-    int size = 0;
-    for (int i = 0; i < fields.getNumArrayEntries(); i++) {
-      final Map.Entry<FieldDescriptorType, Object> entry =
-          fields.getArrayEntryAt(i);
-      size += computeFieldSize(entry.getKey(), entry.getValue());
-    }
-    for (final Map.Entry<FieldDescriptorType, Object> entry :
-         fields.getOverflowEntries()) {
-      size += computeFieldSize(entry.getKey(), entry.getValue());
-    }
-    return size;
-  }
-
-  /**
-   * Like {@link #getSerializedSize} but uses MessageSet wire format.
-   */
-  public int getMessageSetSerializedSize() {
-    int size = 0;
-    for (int i = 0; i < fields.getNumArrayEntries(); i++) {
-      size += getMessageSetSerializedSize(fields.getArrayEntryAt(i));
-    }
-    for (final Map.Entry<FieldDescriptorType, Object> entry :
-             fields.getOverflowEntries()) {
-      size += getMessageSetSerializedSize(entry);
-    }
-    return size;
-  }
-
-  private int getMessageSetSerializedSize(
-      final Map.Entry<FieldDescriptorType, Object> entry) {
-    final FieldDescriptorType descriptor = entry.getKey();
-    Object value = entry.getValue();
-    if (descriptor.getLiteJavaType() == WireFormat.JavaType.MESSAGE
-        && !descriptor.isRepeated() && !descriptor.isPacked()) {
-      if (value instanceof LazyField) {
-        return CodedOutputStream.computeLazyFieldMessageSetExtensionSize(
-            entry.getKey().getNumber(), (LazyField) value);
-      } else {
-        return CodedOutputStream.computeMessageSetExtensionSize(
-            entry.getKey().getNumber(), (MessageLite) value);
-      }
-    } else {
-      return computeFieldSize(descriptor, value);
-    }
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * single tag/value pair of arbitrary type.
-   *
-   * @param type   The field's type.
-   * @param number The field's number.
-   * @param value  Object representing the field's value.  Must be of the exact
-   *               type which would be returned by
-   *               {@link Message#getField(Descriptors.FieldDescriptor)} for
-   *               this field.
-   */
-  static int computeElementSize(
-      final WireFormat.FieldType type, final int number, final Object value) {
-    int tagSize = CodedOutputStream.computeTagSize(number);
-    if (type == WireFormat.FieldType.GROUP) {
-      // Only count the end group tag for proto2 messages as for proto1 the end
-      // group tag will be counted as a part of getSerializedSize().
-        tagSize *= 2;
-    }
-    return tagSize + computeElementSizeNoTag(type, value);
-  }
-
-  /**
-   * Compute the number of bytes that would be needed to encode a
-   * particular value of arbitrary type, excluding tag.
-   *
-   * @param type   The field's type.
-   * @param value  Object representing the field's value.  Must be of the exact
-   *               type which would be returned by
-   *               {@link Message#getField(Descriptors.FieldDescriptor)} for
-   *               this field.
-   */
-  static int computeElementSizeNoTag(
-      final WireFormat.FieldType type, final Object value) {
-    switch (type) {
-      // Note:  Minor violation of 80-char limit rule here because this would
-      //   actually be harder to read if we wrapped the lines.
-      case DOUBLE  : return CodedOutputStream.computeDoubleSizeNoTag  ((Double     )value);
-      case FLOAT   : return CodedOutputStream.computeFloatSizeNoTag   ((Float      )value);
-      case INT64   : return CodedOutputStream.computeInt64SizeNoTag   ((Long       )value);
-      case UINT64  : return CodedOutputStream.computeUInt64SizeNoTag  ((Long       )value);
-      case INT32   : return CodedOutputStream.computeInt32SizeNoTag   ((Integer    )value);
-      case FIXED64 : return CodedOutputStream.computeFixed64SizeNoTag ((Long       )value);
-      case FIXED32 : return CodedOutputStream.computeFixed32SizeNoTag ((Integer    )value);
-      case BOOL    : return CodedOutputStream.computeBoolSizeNoTag    ((Boolean    )value);
-      case GROUP   : return CodedOutputStream.computeGroupSizeNoTag   ((MessageLite)value);
-      case BYTES   :
-        if (value instanceof ByteString) {
-          return CodedOutputStream.computeBytesSizeNoTag((ByteString) value);
-        } else {
-          return CodedOutputStream.computeByteArraySizeNoTag((byte[]) value);
-        }
-      case STRING  :
-        if (value instanceof ByteString) {
-          return CodedOutputStream.computeBytesSizeNoTag((ByteString) value);
-        } else {
-          return CodedOutputStream.computeStringSizeNoTag((String) value);
-        }
-      case UINT32  : return CodedOutputStream.computeUInt32SizeNoTag  ((Integer    )value);
-      case SFIXED32: return CodedOutputStream.computeSFixed32SizeNoTag((Integer    )value);
-      case SFIXED64: return CodedOutputStream.computeSFixed64SizeNoTag((Long       )value);
-      case SINT32  : return CodedOutputStream.computeSInt32SizeNoTag  ((Integer    )value);
-      case SINT64  : return CodedOutputStream.computeSInt64SizeNoTag  ((Long       )value);
-
-      case MESSAGE:
-        if (value instanceof LazyField) {
-          return CodedOutputStream.computeLazyFieldSizeNoTag((LazyField) value);
-        } else {
-          return CodedOutputStream.computeMessageSizeNoTag((MessageLite) value);
-        }
-
-      case ENUM:
-        if (value instanceof Internal.EnumLite) {
-          return CodedOutputStream.computeEnumSizeNoTag(
-              ((Internal.EnumLite) value).getNumber());
-        } else {
-          return CodedOutputStream.computeEnumSizeNoTag((Integer) value);
-        }
-    }
-
-    throw new RuntimeException(
-      "There is no way to get here, but the compiler thinks otherwise.");
-  }
-
-  /**
-   * Compute the number of bytes needed to encode a particular field.
-   */
-  public static int computeFieldSize(final FieldDescriptorLite<?> descriptor,
-                                     final Object value) {
-    WireFormat.FieldType type = descriptor.getLiteType();
-    int number = descriptor.getNumber();
-    if (descriptor.isRepeated()) {
-      if (descriptor.isPacked()) {
-        int dataSize = 0;
-        for (final Object element : (List<?>)value) {
-          dataSize += computeElementSizeNoTag(type, element);
-        }
-        return dataSize +
-            CodedOutputStream.computeTagSize(number) +
-            CodedOutputStream.computeRawVarint32Size(dataSize);
-      } else {
-        int size = 0;
-        for (final Object element : (List<?>)value) {
-          size += computeElementSize(type, number, element);
-        }
-        return size;
-      }
-    } else {
-      return computeElementSize(type, number, value);
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatArrayList.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatArrayList.java
deleted file mode 100644
index 31f6600..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/FloatArrayList.java
+++ /dev/null
@@ -1,272 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.FloatList;
-
-import java.util.Arrays;
-import java.util.Collection;
-import java.util.RandomAccess;
-
-/**
- * An implementation of {@link FloatList} on top of a primitive array.
- *
- * @author dweis@google.com (Daniel Weis)
- */
-final class FloatArrayList
-    extends AbstractProtobufList<Float>
-    implements FloatList, RandomAccess {
-
-  private static final FloatArrayList EMPTY_LIST = new FloatArrayList();
-  static {
-    EMPTY_LIST.makeImmutable();
-  }
-
-  public static FloatArrayList emptyList() {
-    return EMPTY_LIST;
-  }
-
-  /**
-   * The backing store for the list.
-   */
-  private float[] array;
-
-  /**
-   * The size of the list distinct from the length of the array. That is, it is the number of
-   * elements set in the list.
-   */
-  private int size;
-
-  /**
-   * Constructs a new mutable {@code FloatArrayList} with default capacity.
-   */
-  FloatArrayList() {
-    this(new float[DEFAULT_CAPACITY], 0);
-  }
-
-  /**
-   * Constructs a new mutable {@code FloatArrayList}
-   * containing the same elements as {@code other}.
-   */
-  private FloatArrayList(float[] other, int size) {
-    array = other;
-    this.size = size;
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (this == o) {
-      return true;
-    }
-    if (!(o instanceof FloatArrayList)) {
-      return super.equals(o);
-    }
-    FloatArrayList other = (FloatArrayList) o;
-    if (size != other.size) {
-      return false;
-    }
-
-    final float[] arr = other.array;
-    for (int i = 0; i < size; i++) {
-      if (array[i] != arr[i]) {
-        return false;
-      }
-    }
-
-    return true;
-  }
-
-  @Override
-  public int hashCode() {
-    int result = 1;
-    for (int i = 0; i < size; i++) {
-      result = (31 * result) + Float.floatToIntBits(array[i]);
-    }
-    return result;
-  }
-
-  @Override
-  public FloatList mutableCopyWithCapacity(int capacity) {
-    if (capacity < size) {
-      throw new IllegalArgumentException();
-    }
-    return new FloatArrayList(Arrays.copyOf(array, capacity), size);
-  }
-
-  @Override
-  public Float get(int index) {
-    return getFloat(index);
-  }
-
-  @Override
-  public float getFloat(int index) {
-    ensureIndexInRange(index);
-    return array[index];
-  }
-
-  @Override
-  public int size() {
-    return size;
-  }
-
-  @Override
-  public Float set(int index, Float element) {
-    return setFloat(index, element);
-  }
-
-  @Override
-  public float setFloat(int index, float element) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    float previousValue = array[index];
-    array[index] = element;
-    return previousValue;
-  }
-
-  @Override
-  public void add(int index, Float element) {
-    addFloat(index, element);
-  }
-
-  /**
-   * Like {@link #add(Float)} but more efficient in that it doesn't box the element.
-   */
-  @Override
-  public void addFloat(float element) {
-    addFloat(size, element);
-  }
-
-  /**
-   * Like {@link #add(int, Float)} but more efficient in that it doesn't box the element.
-   */
-  private void addFloat(int index, float element) {
-    ensureIsMutable();
-    if (index < 0 || index > size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-
-    if (size < array.length) {
-      // Shift everything over to make room
-      System.arraycopy(array, index, array, index + 1, size - index);
-    } else {
-      // Resize to 1.5x the size
-      int length = ((size * 3) / 2) + 1;
-      float[] newArray = new float[length];
-
-      // Copy the first part directly
-      System.arraycopy(array, 0, newArray, 0, index);
-
-      // Copy the rest shifted over by one to make room
-      System.arraycopy(array, index, newArray, index + 1, size - index);
-      array = newArray;
-    }
-
-    array[index] = element;
-    size++;
-    modCount++;
-  }
-
-  @Override
-  public boolean addAll(Collection<? extends Float> collection) {
-    ensureIsMutable();
-
-    if (collection == null) {
-      throw new NullPointerException();
-    }
-
-    // We specialize when adding another FloatArrayList to avoid boxing elements.
-    if (!(collection instanceof FloatArrayList)) {
-      return super.addAll(collection);
-    }
-
-    FloatArrayList list = (FloatArrayList) collection;
-    if (list.size == 0) {
-      return false;
-    }
-
-    int overflow = Integer.MAX_VALUE - size;
-    if (overflow < list.size) {
-      // We can't actually represent a list this large.
-      throw new OutOfMemoryError();
-    }
-
-    int newSize = size + list.size;
-    if (newSize > array.length) {
-      array = Arrays.copyOf(array, newSize);
-    }
-
-    System.arraycopy(list.array, 0, array, size, list.size);
-    size = newSize;
-    modCount++;
-    return true;
-  }
-
-  @Override
-  public boolean remove(Object o) {
-    ensureIsMutable();
-    for (int i = 0; i < size; i++) {
-      if (o.equals(array[i])) {
-        System.arraycopy(array, i + 1, array, i, size - i);
-        size--;
-        modCount++;
-        return true;
-      }
-    }
-    return false;
-  }
-
-  @Override
-  public Float remove(int index) {
-    ensureIsMutable();
-    ensureIndexInRange(index);
-    float value = array[index];
-    System.arraycopy(array, index + 1, array, index, size - index);
-    size--;
-    modCount++;
-    return value;
-  }
-
-  /**
-   * Ensures that the provided {@code index} is within the range of {@code [0, size]}. Throws an
-   * {@link IndexOutOfBoundsException} if it is not.
-   *
-   * @param index the index to verify is in range
-   */
-  private void ensureIndexInRange(int index) {
-    if (index < 0 || index >= size) {
-      throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index));
-    }
-  }
-
-  private String makeOutOfBoundsExceptionMessage(int index) {
-    return "Index:" + index + ", Size:" + size;
-  }
-}


[13/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseInfoTree.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseInfoTree.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseInfoTree.java
deleted file mode 100644
index cd08eab..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseInfoTree.java
+++ /dev/null
@@ -1,226 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor;
-
-import java.util.ArrayList;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-import java.util.Map.Entry;
-
-
-/**
- * Data structure which is populated with the locations of each field value parsed from the text.
- *
- * <p>The locations of primary fields values are retrieved by {@code getLocation} or
- * {@code getLocations}.  The locations of sub message values are within nested
- * {@code TextFormatParseInfoTree}s and are retrieve by {@code getNestedTree} or
- * {@code getNestedTrees}.
- *
- * <p>The {@code TextFormatParseInfoTree} is created by a Builder.
- */
-public class TextFormatParseInfoTree {
-
-  // Defines a mapping between each field's descriptor to the list of locations where
-  // its value(s) were was encountered.
-  private Map<FieldDescriptor, List<TextFormatParseLocation>> locationsFromField;
-
-  // Defines a mapping between a field's descriptor to a list of TextFormatParseInfoTrees for
-  // sub message location information.
-  Map<FieldDescriptor, List<TextFormatParseInfoTree>> subtreesFromField;
-
-  /**
-   * Construct a {@code TextFormatParseInfoTree}.
-   *
-   * @param locationsFromField a map of fields to location in the source code
-   * @param subtreeBuildersFromField a map of fields to parse tree location information builders
-   */
-  private TextFormatParseInfoTree(
-      Map<FieldDescriptor, List<TextFormatParseLocation>> locationsFromField,
-      Map<FieldDescriptor, List<TextFormatParseInfoTree.Builder>> subtreeBuildersFromField) {
-
-    // The maps are unmodifiable.  The values in the maps are unmodifiable.
-    Map<FieldDescriptor, List<TextFormatParseLocation>> locs =
-        new HashMap<FieldDescriptor, List<TextFormatParseLocation>>();
-    for (Entry<FieldDescriptor, List<TextFormatParseLocation>> kv : locationsFromField.entrySet()) {
-      locs.put(kv.getKey(), Collections.unmodifiableList(kv.getValue()));
-    }
-    this.locationsFromField = Collections.unmodifiableMap(locs);
-
-    Map<FieldDescriptor, List<TextFormatParseInfoTree>> subs =
-        new HashMap<FieldDescriptor, List<TextFormatParseInfoTree>>();
-    for (Entry<FieldDescriptor, List<Builder>> kv : subtreeBuildersFromField.entrySet()) {
-      List<TextFormatParseInfoTree> submessagesOfField = new ArrayList<TextFormatParseInfoTree>();
-      for (Builder subBuilder : kv.getValue()) {
-        submessagesOfField.add(subBuilder.build());
-      }
-      subs.put(kv.getKey(), Collections.unmodifiableList(submessagesOfField));
-    }
-    this.subtreesFromField = Collections.unmodifiableMap(subs);
-  }
-
- /**
-  * Retrieve all the locations of a field.
-  *
-  * @param fieldDescriptor the the @{link FieldDescriptor} of the desired field
-  * @return a list of the locations of values of the field.  If there are not values
-  *         or the field doesn't exist, an empty list is returned.
-  */
-  public List<TextFormatParseLocation> getLocations(final FieldDescriptor fieldDescriptor) {
-    List<TextFormatParseLocation> result = locationsFromField.get(fieldDescriptor);
-    return (result == null) ? Collections.<TextFormatParseLocation>emptyList() : result;
-  }
-
-  /**
-   * Get the location in the source of a field's value.
-   *
-   * <p>Returns the {@link TextFormatParseLocation} for index-th value of the field in the parsed
-   * text.
-   *
-   * @param fieldDescriptor the @{link FieldDescriptor} of the desired field
-   * @param index the index of the value.
-   * @return the {@link TextFormatParseLocation} of the value
-   * @throws IllegalArgumentException index is out of range
-   */
-  public TextFormatParseLocation getLocation(final FieldDescriptor fieldDescriptor, int index) {
-    return getFromList(getLocations(fieldDescriptor), index, fieldDescriptor);
-  }
-
-  /**
-   * Retrieve a list of all the location information trees for a sub message field.
-   *
-   * @param fieldDescriptor the @{link FieldDescriptor} of the desired field
-   * @return A list of {@link TextFormatParseInfoTree}
-   */
-  public List<TextFormatParseInfoTree> getNestedTrees(final FieldDescriptor fieldDescriptor) {
-    List<TextFormatParseInfoTree> result = subtreesFromField.get(fieldDescriptor);
-    return result == null ? Collections.<TextFormatParseInfoTree>emptyList() : result;
-  }
-
-  /**
-   * Returns the parse info tree for the given field, which must be a message type.
-   *
-   * @param fieldDescriptor the @{link FieldDescriptor} of the desired sub message
-   * @param index the index of message value.
-   * @return the {@code ParseInfoTree} of the message value. {@code null} is returned if the field
-   *         doesn't exist or the index is out of range.
-   * @throws IllegalArgumentException if index is out of range
-   */
-  public TextFormatParseInfoTree getNestedTree(final FieldDescriptor fieldDescriptor, int index) {
-    return getFromList(getNestedTrees(fieldDescriptor), index, fieldDescriptor);
-  }
-
-  /**
-   * Create a builder for a {@code ParseInfoTree}.
-   *
-   * @return the builder
-   */
-  public static Builder builder() {
-    return new Builder();
-  }
-
-  private static <T> T getFromList(List<T> list, int index, FieldDescriptor fieldDescriptor) {
-    if (index >= list.size() || index < 0)  {
-      throw new IllegalArgumentException(String.format("Illegal index field: %s, index %d",
-          fieldDescriptor == null ? "<null>" : fieldDescriptor.getName(), index));
-    }
-    return list.get(index);
-  }
-
-  /**
-   * Builder for a {@link TextFormatParseInfoTree}.
-   */
-  public static class Builder {
-
-    private Map<FieldDescriptor, List<TextFormatParseLocation>> locationsFromField;
-
-    // Defines a mapping between a field's descriptor to a list of ParseInfoTrees builders for
-    // sub message location information.
-    private Map<FieldDescriptor, List<Builder>> subtreeBuildersFromField;
-
-     /**
-     * Create a root level {@ParseInfoTree} builder.
-     */
-    private Builder() {
-      locationsFromField = new HashMap<FieldDescriptor, List<TextFormatParseLocation>>();
-      subtreeBuildersFromField = new HashMap<FieldDescriptor, List<Builder>>();
-    }
-
-    /**
-     * Record the starting location of a single value for a field.
-     *
-     * @param fieldDescriptor the field
-     * @param location source code location information
-     */
-    public Builder setLocation(
-        final FieldDescriptor fieldDescriptor, TextFormatParseLocation location) {
-      List<TextFormatParseLocation> fieldLocations = locationsFromField.get(fieldDescriptor);
-      if (fieldLocations == null) {
-        fieldLocations = new ArrayList<TextFormatParseLocation>();
-        locationsFromField.put(fieldDescriptor, fieldLocations);
-      }
-      fieldLocations.add(location);
-      return this;
-    }
-
-    /**
-     * Set for a sub message.
-     *
-     * <p>A new builder is created for a sub message. The builder that is returned is a new builder.
-     * The return is <em>not</em> the invoked {@code builder.getBuilderForSubMessageField}.
-     *
-     * @param fieldDescriptor the field whose value is the submessage
-     * @return a new Builder for the sub message
-     */
-    public Builder getBuilderForSubMessageField(final FieldDescriptor fieldDescriptor) {
-      List<Builder> submessageBuilders = subtreeBuildersFromField.get(fieldDescriptor);
-      if (submessageBuilders == null) {
-        submessageBuilders = new ArrayList<Builder>();
-        subtreeBuildersFromField.put(fieldDescriptor, submessageBuilders);
-      }
-      Builder subtreeBuilder = new Builder();
-      submessageBuilders.add(subtreeBuilder);
-      return subtreeBuilder;
-    }
-
-    /**
-     * Build the {@code TextFormatParseInfoTree}.
-     *
-     * @return the {@code TextFormatParseInfoTree}
-     */
-    public TextFormatParseInfoTree build() {
-      return new TextFormatParseInfoTree(locationsFromField, subtreeBuildersFromField);
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseLocation.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseLocation.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseLocation.java
deleted file mode 100644
index 89ca9d2..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TextFormatParseLocation.java
+++ /dev/null
@@ -1,104 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.Arrays;
-
-/**
- * A location in the source code.
- *
- * <p>A location is the starting line number and starting column number.
- */
-public final class TextFormatParseLocation {
-
-  /**
-   * The empty location.
-   */
-  public static final TextFormatParseLocation EMPTY = new TextFormatParseLocation(-1, -1);
-
-  /**
-   * Create a location.
-   *
-   * @param line the starting line number
-   * @param column the starting column number
-   * @return a {@code ParseLocation}
-   */
-  static TextFormatParseLocation create(int line, int column) {
-    if (line == -1 && column == -1) {
-      return EMPTY;
-    }
-    if (line < 0 || column < 0) {
-      throw new IllegalArgumentException(
-          String.format("line and column values must be >= 0: line %d, column: %d", line, column));
-    }
-    return new TextFormatParseLocation(line, column);
-  }
-
-  private final int line;
-  private final int column;
-
-  private TextFormatParseLocation(int line, int column) {
-    this.line = line;
-    this.column = column;
-  }
-
-  public int getLine() {
-    return line;
-  }
-
-  public int getColumn() {
-    return column;
-  }
-
-  @Override
-  public String toString() {
-    return String.format("ParseLocation{line=%d, column=%d}", line, column);
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (o == this) {
-      return true;
-    }
-    if (!(o instanceof TextFormatParseLocation)) {
-      return false;
-    }
-    TextFormatParseLocation that = (TextFormatParseLocation) o;
-    return (this.line == that.getLine())
-         && (this.column == that.getColumn());
-  }
-
-  @Override
-  public int hashCode() {
-    int[] values = {line, column};
-    return Arrays.hashCode(values);
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Timestamp.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Timestamp.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Timestamp.java
deleted file mode 100644
index 0023d50..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Timestamp.java
+++ /dev/null
@@ -1,616 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/timestamp.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * A Timestamp represents a point in time independent of any time zone
- * or calendar, represented as seconds and fractions of seconds at
- * nanosecond resolution in UTC Epoch time. It is encoded using the
- * Proleptic Gregorian Calendar which extends the Gregorian calendar
- * backwards to year one. It is encoded assuming all minutes are 60
- * seconds long, i.e. leap seconds are "smeared" so that no leap second
- * table is needed for interpretation. Range is from
- * 0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z.
- * By restricting to that range, we ensure that we can convert to
- * and from  RFC 3339 date strings.
- * See [https://www.ietf.org/rfc/rfc3339.txt](https://www.ietf.org/rfc/rfc3339.txt).
- * Example 1: Compute Timestamp from POSIX `time()`.
- *     Timestamp timestamp;
- *     timestamp.set_seconds(time(NULL));
- *     timestamp.set_nanos(0);
- * Example 2: Compute Timestamp from POSIX `gettimeofday()`.
- *     struct timeval tv;
- *     gettimeofday(&amp;tv, NULL);
- *     Timestamp timestamp;
- *     timestamp.set_seconds(tv.tv_sec);
- *     timestamp.set_nanos(tv.tv_usec * 1000);
- * Example 3: Compute Timestamp from Win32 `GetSystemTimeAsFileTime()`.
- *     FILETIME ft;
- *     GetSystemTimeAsFileTime(&amp;ft);
- *     UINT64 ticks = (((UINT64)ft.dwHighDateTime) &lt;&lt; 32) | ft.dwLowDateTime;
- *     // A Windows tick is 100 nanoseconds. Windows epoch 1601-01-01T00:00:00Z
- *     // is 11644473600 seconds before Unix epoch 1970-01-01T00:00:00Z.
- *     Timestamp timestamp;
- *     timestamp.set_seconds((INT64) ((ticks / 10000000) - 11644473600LL));
- *     timestamp.set_nanos((INT32) ((ticks % 10000000) * 100));
- * Example 4: Compute Timestamp from Java `System.currentTimeMillis()`.
- *     long millis = System.currentTimeMillis();
- *     Timestamp timestamp = Timestamp.newBuilder().setSeconds(millis / 1000)
- *         .setNanos((int) ((millis % 1000) * 1000000)).build();
- * Example 5: Compute Timestamp from current time in Python.
- *     timestamp = Timestamp()
- *     timestamp.GetCurrentTime()
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Timestamp}
- */
-public  final class Timestamp extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Timestamp)
-    TimestampOrBuilder {
-  // Use Timestamp.newBuilder() to construct.
-  private Timestamp(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Timestamp() {
-    seconds_ = 0L;
-    nanos_ = 0;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Timestamp(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-
-            seconds_ = input.readInt64();
-            break;
-          }
-          case 16: {
-
-            nanos_ = input.readInt32();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.Builder.class);
-  }
-
-  public static final int SECONDS_FIELD_NUMBER = 1;
-  private long seconds_;
-  /**
-   * <pre>
-   * Represents seconds of UTC time since Unix epoch
-   * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
-   * 9999-12-31T23:59:59Z inclusive.
-   * </pre>
-   *
-   * <code>int64 seconds = 1;</code>
-   */
-  public long getSeconds() {
-    return seconds_;
-  }
-
-  public static final int NANOS_FIELD_NUMBER = 2;
-  private int nanos_;
-  /**
-   * <pre>
-   * Non-negative fractions of a second at nanosecond resolution. Negative
-   * second values with fractions must still have non-negative nanos values
-   * that count forward in time. Must be from 0 to 999,999,999
-   * inclusive.
-   * </pre>
-   *
-   * <code>int32 nanos = 2;</code>
-   */
-  public int getNanos() {
-    return nanos_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (seconds_ != 0L) {
-      output.writeInt64(1, seconds_);
-    }
-    if (nanos_ != 0) {
-      output.writeInt32(2, nanos_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (seconds_ != 0L) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt64Size(1, seconds_);
-    }
-    if (nanos_ != 0) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeInt32Size(2, nanos_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp) obj;
-
-    boolean result = true;
-    result = result && (getSeconds()
-        == other.getSeconds());
-    result = result && (getNanos()
-        == other.getNanos());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + SECONDS_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
-        getSeconds());
-    hash = (37 * hash) + NANOS_FIELD_NUMBER;
-    hash = (53 * hash) + getNanos();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * A Timestamp represents a point in time independent of any time zone
-   * or calendar, represented as seconds and fractions of seconds at
-   * nanosecond resolution in UTC Epoch time. It is encoded using the
-   * Proleptic Gregorian Calendar which extends the Gregorian calendar
-   * backwards to year one. It is encoded assuming all minutes are 60
-   * seconds long, i.e. leap seconds are "smeared" so that no leap second
-   * table is needed for interpretation. Range is from
-   * 0001-01-01T00:00:00Z to 9999-12-31T23:59:59.999999999Z.
-   * By restricting to that range, we ensure that we can convert to
-   * and from  RFC 3339 date strings.
-   * See [https://www.ietf.org/rfc/rfc3339.txt](https://www.ietf.org/rfc/rfc3339.txt).
-   * Example 1: Compute Timestamp from POSIX `time()`.
-   *     Timestamp timestamp;
-   *     timestamp.set_seconds(time(NULL));
-   *     timestamp.set_nanos(0);
-   * Example 2: Compute Timestamp from POSIX `gettimeofday()`.
-   *     struct timeval tv;
-   *     gettimeofday(&amp;tv, NULL);
-   *     Timestamp timestamp;
-   *     timestamp.set_seconds(tv.tv_sec);
-   *     timestamp.set_nanos(tv.tv_usec * 1000);
-   * Example 3: Compute Timestamp from Win32 `GetSystemTimeAsFileTime()`.
-   *     FILETIME ft;
-   *     GetSystemTimeAsFileTime(&amp;ft);
-   *     UINT64 ticks = (((UINT64)ft.dwHighDateTime) &lt;&lt; 32) | ft.dwLowDateTime;
-   *     // A Windows tick is 100 nanoseconds. Windows epoch 1601-01-01T00:00:00Z
-   *     // is 11644473600 seconds before Unix epoch 1970-01-01T00:00:00Z.
-   *     Timestamp timestamp;
-   *     timestamp.set_seconds((INT64) ((ticks / 10000000) - 11644473600LL));
-   *     timestamp.set_nanos((INT32) ((ticks % 10000000) * 100));
-   * Example 4: Compute Timestamp from Java `System.currentTimeMillis()`.
-   *     long millis = System.currentTimeMillis();
-   *     Timestamp timestamp = Timestamp.newBuilder().setSeconds(millis / 1000)
-   *         .setNanos((int) ((millis % 1000) * 1000000)).build();
-   * Example 5: Compute Timestamp from current time in Python.
-   *     timestamp = Timestamp()
-   *     timestamp.GetCurrentTime()
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Timestamp}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Timestamp)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      seconds_ = 0L;
-
-      nanos_ = 0;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.TimestampProto.internal_static_google_protobuf_Timestamp_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp(this);
-      result.seconds_ = seconds_;
-      result.nanos_ = nanos_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp.getDefaultInstance()) return this;
-      if (other.getSeconds() != 0L) {
-        setSeconds(other.getSeconds());
-      }
-      if (other.getNanos() != 0) {
-        setNanos(other.getNanos());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private long seconds_ ;
-    /**
-     * <pre>
-     * Represents seconds of UTC time since Unix epoch
-     * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
-     * 9999-12-31T23:59:59Z inclusive.
-     * </pre>
-     *
-     * <code>int64 seconds = 1;</code>
-     */
-    public long getSeconds() {
-      return seconds_;
-    }
-    /**
-     * <pre>
-     * Represents seconds of UTC time since Unix epoch
-     * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
-     * 9999-12-31T23:59:59Z inclusive.
-     * </pre>
-     *
-     * <code>int64 seconds = 1;</code>
-     */
-    public Builder setSeconds(long value) {
-      
-      seconds_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Represents seconds of UTC time since Unix epoch
-     * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
-     * 9999-12-31T23:59:59Z inclusive.
-     * </pre>
-     *
-     * <code>int64 seconds = 1;</code>
-     */
-    public Builder clearSeconds() {
-      
-      seconds_ = 0L;
-      onChanged();
-      return this;
-    }
-
-    private int nanos_ ;
-    /**
-     * <pre>
-     * Non-negative fractions of a second at nanosecond resolution. Negative
-     * second values with fractions must still have non-negative nanos values
-     * that count forward in time. Must be from 0 to 999,999,999
-     * inclusive.
-     * </pre>
-     *
-     * <code>int32 nanos = 2;</code>
-     */
-    public int getNanos() {
-      return nanos_;
-    }
-    /**
-     * <pre>
-     * Non-negative fractions of a second at nanosecond resolution. Negative
-     * second values with fractions must still have non-negative nanos values
-     * that count forward in time. Must be from 0 to 999,999,999
-     * inclusive.
-     * </pre>
-     *
-     * <code>int32 nanos = 2;</code>
-     */
-    public Builder setNanos(int value) {
-      
-      nanos_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Non-negative fractions of a second at nanosecond resolution. Negative
-     * second values with fractions must still have non-negative nanos values
-     * that count forward in time. Must be from 0 to 999,999,999
-     * inclusive.
-     * </pre>
-     *
-     * <code>int32 nanos = 2;</code>
-     */
-    public Builder clearNanos() {
-      
-      nanos_ = 0;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Timestamp)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Timestamp)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Timestamp>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Timestamp>() {
-    public Timestamp parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Timestamp(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Timestamp> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Timestamp> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Timestamp getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampOrBuilder.java
deleted file mode 100644
index fca3c16..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampOrBuilder.java
+++ /dev/null
@@ -1,32 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/timestamp.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface TimestampOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Timestamp)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * Represents seconds of UTC time since Unix epoch
-   * 1970-01-01T00:00:00Z. Must be from 0001-01-01T00:00:00Z to
-   * 9999-12-31T23:59:59Z inclusive.
-   * </pre>
-   *
-   * <code>int64 seconds = 1;</code>
-   */
-  long getSeconds();
-
-  /**
-   * <pre>
-   * Non-negative fractions of a second at nanosecond resolution. Negative
-   * second values with fractions must still have non-negative nanos values
-   * that count forward in time. Must be from 0 to 999,999,999
-   * inclusive.
-   * </pre>
-   *
-   * <code>int32 nanos = 2;</code>
-   */
-  int getNanos();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampProto.java
deleted file mode 100644
index c62c4ce..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/TimestampProto.java
+++ /dev/null
@@ -1,59 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/timestamp.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class TimestampProto {
-  private TimestampProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Timestamp_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Timestamp_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\037google/protobuf/timestamp.proto\022\017googl" +
-      "e.protobuf\"+\n\tTimestamp\022\017\n\007seconds\030\001 \001(\003" +
-      "\022\r\n\005nanos\030\002 \001(\005B~\n\023com.google.protobufB\016" +
-      "TimestampProtoP\001Z+github.com/golang/prot" +
-      "obuf/ptypes/timestamp\370\001\001\242\002\003GPB\252\002\036Google." +
-      "Protobuf.WellKnownTypesb\006proto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_google_protobuf_Timestamp_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_Timestamp_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Timestamp_descriptor,
-        new java.lang.String[] { "Seconds", "Nanos", });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[51/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
HBASE-17056 Remove checked in PB generated files

Selective add of dependency on hbase-thirdparty jars.
Update to READMEs on how protobuf is done (and update to refguide).
Removed all checked in generated protobuf files. They are generated
on the fly now as part of mainline build.


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/ee70b1d2
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/ee70b1d2
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/ee70b1d2

Branch: refs/heads/branch-2
Commit: ee70b1d2e05f4b007ffbf7698b192b44c986c8ae
Parents: f923342
Author: Michael Stack <st...@apache.org>
Authored: Wed Jul 5 12:06:29 2017 -0700
Committer: Michael Stack <st...@apache.org>
Committed: Wed Aug 2 09:42:38 2017 -0700

----------------------------------------------------------------------
 hbase-client/pom.xml                            |     4 +
 hbase-endpoint/README.txt                       |    25 +-
 hbase-endpoint/pom.xml                          |   258 +-
 .../example/generated/BulkDeleteProtos.java     |  1792 -
 .../generated/ColumnAggregationProtos.java      |  1277 -
 .../ColumnAggregationWithErrorsProtos.java      |  1290 -
 ...ColumnAggregationWithNullResponseProtos.java |  1283 -
 .../DummyRegionServerEndpointProtos.java        |  1225 -
 .../IncrementCounterProcessorTestProtos.java    |  4059 -
 .../protobuf/generated/AggregateProtos.java     |  2375 -
 .../generated/SecureBulkLoadProtos.java         |  2088 -
 hbase-examples/README.txt                       |    27 +-
 hbase-examples/pom.xml                          |   320 +-
 .../example/generated/ExampleProtos.java        |  1149 -
 hbase-procedure/pom.xml                         |     4 +
 hbase-protocol-shaded/README.txt                |    49 +-
 hbase-protocol-shaded/pom.xml                   |   325 +-
 .../com/google/protobuf/AbstractMessage.java    |   646 -
 .../google/protobuf/AbstractMessageLite.java    |   383 -
 .../com/google/protobuf/AbstractParser.java     |   258 -
 .../google/protobuf/AbstractProtobufList.java   |   180 -
 .../hbase/shaded/com/google/protobuf/Any.java   |   899 -
 .../com/google/protobuf/AnyOrBuilder.java       |    70 -
 .../shaded/com/google/protobuf/AnyProto.java    |    59 -
 .../hbase/shaded/com/google/protobuf/Api.java   |  2473 -
 .../com/google/protobuf/ApiOrBuilder.java       |   258 -
 .../shaded/com/google/protobuf/ApiProto.java    |    98 -
 .../com/google/protobuf/BlockingRpcChannel.java |    51 -
 .../com/google/protobuf/BlockingService.java    |    64 -
 .../shaded/com/google/protobuf/BoolValue.java   |   452 -
 .../com/google/protobuf/BoolValueOrBuilder.java |    18 -
 .../com/google/protobuf/BooleanArrayList.java   |   272 -
 .../com/google/protobuf/ByteBufferWriter.java   |   185 -
 .../shaded/com/google/protobuf/ByteInput.java   |    81 -
 .../google/protobuf/ByteInputByteString.java    |   251 -
 .../shaded/com/google/protobuf/ByteOutput.java  |   116 -
 .../shaded/com/google/protobuf/ByteString.java  |  1565 -
 .../shaded/com/google/protobuf/BytesValue.java  |   454 -
 .../google/protobuf/BytesValueOrBuilder.java    |    18 -
 .../com/google/protobuf/CodedInputStream.java   |  3549 -
 .../com/google/protobuf/CodedOutputStream.java  |  3001 -
 .../com/google/protobuf/DescriptorProtos.java   | 39141 --------
 .../shaded/com/google/protobuf/Descriptors.java |  2547 -
 .../com/google/protobuf/DoubleArrayList.java    |   273 -
 .../shaded/com/google/protobuf/DoubleValue.java |   454 -
 .../google/protobuf/DoubleValueOrBuilder.java   |    18 -
 .../shaded/com/google/protobuf/Duration.java    |   618 -
 .../com/google/protobuf/DurationOrBuilder.java  |    33 -
 .../com/google/protobuf/DurationProto.java      |    59 -
 .../com/google/protobuf/DynamicMessage.java     |   684 -
 .../hbase/shaded/com/google/protobuf/Empty.java |   386 -
 .../com/google/protobuf/EmptyOrBuilder.java     |     9 -
 .../shaded/com/google/protobuf/EmptyProto.java  |    58 -
 .../hbase/shaded/com/google/protobuf/Enum.java  |  1745 -
 .../com/google/protobuf/EnumOrBuilder.java      |   157 -
 .../shaded/com/google/protobuf/EnumValue.java   |  1044 -
 .../com/google/protobuf/EnumValueOrBuilder.java |    80 -
 .../com/google/protobuf/ExperimentalApi.java    |    66 -
 .../shaded/com/google/protobuf/Extension.java   |    86 -
 .../com/google/protobuf/ExtensionLite.java      |    63 -
 .../com/google/protobuf/ExtensionRegistry.java  |   396 -
 .../protobuf/ExtensionRegistryFactory.java      |    95 -
 .../google/protobuf/ExtensionRegistryLite.java  |   227 -
 .../hbase/shaded/com/google/protobuf/Field.java |  2450 -
 .../shaded/com/google/protobuf/FieldMask.java   |   903 -
 .../com/google/protobuf/FieldMaskOrBuilder.java |    44 -
 .../com/google/protobuf/FieldMaskProto.java     |    59 -
 .../com/google/protobuf/FieldOrBuilder.java     |   189 -
 .../shaded/com/google/protobuf/FieldSet.java    |   909 -
 .../com/google/protobuf/FloatArrayList.java     |   272 -
 .../shaded/com/google/protobuf/FloatValue.java  |   454 -
 .../google/protobuf/FloatValueOrBuilder.java    |    18 -
 .../com/google/protobuf/GeneratedMessage.java   |  3047 -
 .../google/protobuf/GeneratedMessageLite.java   |  2419 -
 .../com/google/protobuf/GeneratedMessageV3.java |  2838 -
 .../shaded/com/google/protobuf/Int32Value.java  |   451 -
 .../google/protobuf/Int32ValueOrBuilder.java    |    18 -
 .../shaded/com/google/protobuf/Int64Value.java  |   452 -
 .../google/protobuf/Int64ValueOrBuilder.java    |    18 -
 .../com/google/protobuf/IntArrayList.java       |   272 -
 .../shaded/com/google/protobuf/Internal.java    |   751 -
 .../InvalidProtocolBufferException.java         |   146 -
 .../shaded/com/google/protobuf/LazyField.java   |   154 -
 .../com/google/protobuf/LazyFieldLite.java      |   437 -
 .../google/protobuf/LazyStringArrayList.java    |   423 -
 .../com/google/protobuf/LazyStringList.java     |   174 -
 .../shaded/com/google/protobuf/ListValue.java   |   814 -
 .../com/google/protobuf/ListValueOrBuilder.java |    53 -
 .../com/google/protobuf/LongArrayList.java      |   272 -
 .../shaded/com/google/protobuf/MapEntry.java    |   449 -
 .../com/google/protobuf/MapEntryLite.java       |   226 -
 .../shaded/com/google/protobuf/MapField.java    |   624 -
 .../com/google/protobuf/MapFieldLite.java       |   224 -
 .../shaded/com/google/protobuf/Message.java     |   292 -
 .../shaded/com/google/protobuf/MessageLite.java |   341 -
 .../google/protobuf/MessageLiteOrBuilder.java   |    60 -
 .../google/protobuf/MessageLiteToString.java    |   239 -
 .../com/google/protobuf/MessageOrBuilder.java   |   143 -
 .../com/google/protobuf/MessageReflection.java  |   990 -
 .../shaded/com/google/protobuf/Method.java      |  1541 -
 .../com/google/protobuf/MethodOrBuilder.java    |   142 -
 .../hbase/shaded/com/google/protobuf/Mixin.java |   812 -
 .../com/google/protobuf/MixinOrBuilder.java     |    47 -
 .../com/google/protobuf/MutabilityOracle.java   |    48 -
 .../com/google/protobuf/NioByteString.java      |   291 -
 .../shaded/com/google/protobuf/NullValue.java   |   108 -
 .../shaded/com/google/protobuf/Option.java      |   818 -
 .../com/google/protobuf/OptionOrBuilder.java    |    67 -
 .../shaded/com/google/protobuf/Parser.java      |   272 -
 .../com/google/protobuf/ProtobufArrayList.java  |   105 -
 .../google/protobuf/ProtocolMessageEnum.java    |    59 -
 .../com/google/protobuf/ProtocolStringList.java |    48 -
 .../google/protobuf/RepeatedFieldBuilder.java   |   708 -
 .../google/protobuf/RepeatedFieldBuilderV3.java |   708 -
 .../com/google/protobuf/RopeByteString.java     |   897 -
 .../shaded/com/google/protobuf/RpcCallback.java |    47 -
 .../shaded/com/google/protobuf/RpcChannel.java  |    71 -
 .../com/google/protobuf/RpcController.java      |   118 -
 .../shaded/com/google/protobuf/RpcUtil.java     |   136 -
 .../shaded/com/google/protobuf/Service.java     |   117 -
 .../com/google/protobuf/ServiceException.java   |    52 -
 .../com/google/protobuf/SingleFieldBuilder.java |   241 -
 .../google/protobuf/SingleFieldBuilderV3.java   |   241 -
 .../com/google/protobuf/SmallSortedMap.java     |   673 -
 .../com/google/protobuf/SourceContext.java      |   539 -
 .../google/protobuf/SourceContextOrBuilder.java |    29 -
 .../com/google/protobuf/SourceContextProto.java |    60 -
 .../shaded/com/google/protobuf/StringValue.java |   532 -
 .../google/protobuf/StringValueOrBuilder.java   |    27 -
 .../shaded/com/google/protobuf/Struct.java      |   705 -
 .../com/google/protobuf/StructOrBuilder.java    |    63 -
 .../shaded/com/google/protobuf/StructProto.java |   103 -
 .../shaded/com/google/protobuf/Syntax.java      |   123 -
 .../shaded/com/google/protobuf/TextFormat.java  |  2108 -
 .../com/google/protobuf/TextFormatEscaper.java  |   137 -
 .../protobuf/TextFormatParseInfoTree.java       |   226 -
 .../protobuf/TextFormatParseLocation.java       |   104 -
 .../shaded/com/google/protobuf/Timestamp.java   |   616 -
 .../com/google/protobuf/TimestampOrBuilder.java |    32 -
 .../com/google/protobuf/TimestampProto.java     |    59 -
 .../hbase/shaded/com/google/protobuf/Type.java  |  1967 -
 .../com/google/protobuf/TypeOrBuilder.java      |   192 -
 .../shaded/com/google/protobuf/TypeProto.java   |   141 -
 .../shaded/com/google/protobuf/UInt32Value.java |   451 -
 .../google/protobuf/UInt32ValueOrBuilder.java   |    18 -
 .../shaded/com/google/protobuf/UInt64Value.java |   452 -
 .../google/protobuf/UInt64ValueOrBuilder.java   |    18 -
 .../protobuf/UninitializedMessageException.java |    99 -
 .../com/google/protobuf/UnknownFieldSet.java    |  1042 -
 .../google/protobuf/UnknownFieldSetLite.java    |   432 -
 .../protobuf/UnmodifiableLazyStringList.java    |   210 -
 .../google/protobuf/UnsafeByteOperations.java   |   131 -
 .../shaded/com/google/protobuf/UnsafeUtil.java  |   295 -
 .../hbase/shaded/com/google/protobuf/Utf8.java  |  1764 -
 .../hbase/shaded/com/google/protobuf/Value.java |  1410 -
 .../com/google/protobuf/ValueOrBuilder.java     |    98 -
 .../shaded/com/google/protobuf/WireFormat.java  |   260 -
 .../com/google/protobuf/WrappersProto.java      |   153 -
 .../google/protobuf/compiler/PluginProtos.java  |  5424 --
 .../protobuf/generated/TestProcedureProtos.java |   608 -
 .../ipc/protobuf/generated/TestProtos.java      |  2987 -
 .../generated/TestRpcServiceProtos.java         |   575 -
 .../protobuf/generated/AccessControlProtos.java | 11171 ---
 .../shaded/protobuf/generated/AdminProtos.java  | 28175 ------
 .../shaded/protobuf/generated/BackupProtos.java |  7013 --
 .../shaded/protobuf/generated/CellProtos.java   |  2287 -
 .../shaded/protobuf/generated/ClientProtos.java | 42367 ---------
 .../protobuf/generated/ClusterIdProtos.java     |   668 -
 .../protobuf/generated/ClusterStatusProtos.java | 15191 ----
 .../protobuf/generated/ComparatorProtos.java    |  5719 --
 .../protobuf/generated/EncryptionProtos.java    |   980 -
 .../protobuf/generated/ErrorHandlingProtos.java |  2997 -
 .../shaded/protobuf/generated/FSProtos.java     |  1311 -
 .../shaded/protobuf/generated/FilterProtos.java | 18739 ----
 .../shaded/protobuf/generated/HBaseProtos.java  | 19346 ----
 .../shaded/protobuf/generated/HFileProtos.java  |  2449 -
 .../protobuf/generated/LoadBalancerProtos.java  |   524 -
 .../protobuf/generated/LockServiceProtos.java   |  7595 --
 .../protobuf/generated/MapReduceProtos.java     |  1804 -
 .../generated/MasterProcedureProtos.java        | 32136 -------
 .../shaded/protobuf/generated/MasterProtos.java | 80755 -----------------
 .../protobuf/generated/ProcedureProtos.java     |  7844 --
 .../shaded/protobuf/generated/QuotaProtos.java  | 14525 ---
 .../shaded/protobuf/generated/RPCProtos.java    |  8299 --
 .../generated/RegionNormalizerProtos.java       |   524 -
 .../generated/RegionServerStatusProtos.java     | 11621 ---
 .../protobuf/generated/ReplicationProtos.java   | 12909 ---
 .../protobuf/generated/SnapshotProtos.java      |  6491 --
 .../protobuf/generated/TracingProtos.java       |   628 -
 .../shaded/protobuf/generated/WALProtos.java    | 12308 ---
 .../protobuf/generated/ZooKeeperProtos.java     |  4717 -
 hbase-protocol/README.txt                       |    16 +-
 hbase-protocol/pom.xml                          |   266 +-
 .../protobuf/generated/PingProtos.java          |  4890 -
 .../protobuf/generated/TestProcedureProtos.java |   530 -
 .../ipc/protobuf/generated/TestProtos.java      |  2778 -
 .../generated/TestRpcServiceProtos.java         |   568 -
 .../protobuf/generated/AccessControlProtos.java | 10656 ---
 .../hbase/protobuf/generated/AdminProtos.java   | 24326 -----
 .../generated/AuthenticationProtos.java         |  4659 -
 .../hbase/protobuf/generated/CellProtos.java    |  2243 -
 .../hbase/protobuf/generated/ClientProtos.java  | 41410 ---------
 .../protobuf/generated/ClusterIdProtos.java     |   628 -
 .../protobuf/generated/ClusterStatusProtos.java | 14975 ---
 .../protobuf/generated/ComparatorProtos.java    |  5419 --
 .../protobuf/generated/EncryptionProtos.java    |   954 -
 .../protobuf/generated/ErrorHandlingProtos.java |  2895 -
 .../hbase/protobuf/generated/FSProtos.java      |  1235 -
 .../hbase/protobuf/generated/FilterProtos.java  | 17809 ----
 .../hbase/protobuf/generated/HBaseProtos.java   | 19326 ----
 .../hbase/protobuf/generated/HFileProtos.java   |  2403 -
 .../protobuf/generated/LoadBalancerProtos.java  |   485 -
 .../protobuf/generated/MapReduceProtos.java     |  1737 -
 .../generated/MultiRowMutationProtos.java       |  2458 -
 .../hbase/protobuf/generated/QuotaProtos.java   |  7342 --
 .../hbase/protobuf/generated/RPCProtos.java     |  6299 --
 .../protobuf/generated/RowProcessorProtos.java  |  1701 -
 .../protobuf/generated/SnapshotProtos.java      |  4789 -
 .../hbase/protobuf/generated/TableProtos.java   |   607 -
 .../hbase/protobuf/generated/TracingProtos.java |   591 -
 .../generated/VisibilityLabelsProtos.java       |  6649 --
 .../hbase/protobuf/generated/WALProtos.java     | 12109 ---
 .../protobuf/generated/ZooKeeperProtos.java     |  8412 --
 hbase-rest/README.txt                           |    21 +-
 hbase-rest/pom.xml                              |    67 +-
 .../rest/protobuf/generated/CellMessage.java    |   731 -
 .../rest/protobuf/generated/CellSetMessage.java |  1521 -
 .../protobuf/generated/ColumnSchemaMessage.java |  1904 -
 .../generated/NamespacePropertiesMessage.java   |  1394 -
 .../protobuf/generated/NamespacesMessage.java   |   547 -
 .../rest/protobuf/generated/ScannerMessage.java |  1578 -
 .../generated/StorageClusterStatusMessage.java  |  3955 -
 .../protobuf/generated/TableInfoMessage.java    |  1802 -
 .../protobuf/generated/TableListMessage.java    |   547 -
 .../protobuf/generated/TableSchemaMessage.java  |  2125 -
 .../rest/protobuf/generated/VersionMessage.java |  1147 -
 hbase-rsgroup/README.txt                        |    29 +-
 hbase-rsgroup/pom.xml                           |    68 +-
 .../protobuf/generated/RSGroupAdminProtos.java  | 13571 ---
 .../hbase/protobuf/generated/RSGroupProtos.java |  1332 -
 hbase-server/pom.xml                            |    12 +
 hbase-spark/README.txt                          |    32 +-
 hbase-spark/pom.xml                             |  1275 +-
 .../protobuf/generated/SparkFilterProtos.java   |  2006 -
 pom.xml                                         |    17 +-
 src/main/asciidoc/_chapters/protobuf.adoc       |    10 +-
 246 files changed, 1223 insertions(+), 742098 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-client/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-client/pom.xml b/hbase-client/pom.xml
index 3b072c4..df94cf9 100644
--- a/hbase-client/pom.xml
+++ b/hbase-client/pom.xml
@@ -87,6 +87,10 @@
   </build>
 
   <dependencies>
+    <dependency>
+      <groupId>org.apache.hbase.thirdparty</groupId>
+      <artifactId>hbase-shaded-protobuf</artifactId>
+    </dependency>
     <!-- Intra-project dependencies -->
     <dependency>
       <groupId>org.apache.hbase</groupId>

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/README.txt
----------------------------------------------------------------------
diff --git a/hbase-endpoint/README.txt b/hbase-endpoint/README.txt
index 861a776..2d2cb4b 100644
--- a/hbase-endpoint/README.txt
+++ b/hbase-endpoint/README.txt
@@ -1,24 +1,13 @@
 ON PROTOBUFS
-This maven module has  protobuf definition files ('.protos') used by hbase
-Coprocessor Endpoints that ship with hbase core including tests. Coprocessor
+This maven module has protobuf definition files ('.protos') used by hbase
+Coprocessor Endpoints that ship with hbase core (including tests). Coprocessor
 Endpoints are meant to be standalone, independent code not reliant on hbase
 internals. They define their Service using protobuf. The protobuf version
 they use can be distinct from that used by HBase internally since HBase started
 shading its protobuf references. Endpoints have no access to the shaded protobuf
-hbase uses. They do have access to the content of hbase-protocol but avoid using
-as much of this as you can as it is liable to change.
+hbase uses. They do have access to the content of hbase-protocol -- the
+.protos found in this module -- but avoid using as much of this as you can as it is
+liable to change.
 
-Generation of java files from protobuf .proto files included here is done apart
-from the build. Run the generation whenever you make changes to the .orotos files
-and then check in the produced java (The reasoning is that change is infrequent
-so why pay the price of generating files anew on each build.
-
-To generate java files from protos run:
-
- $ mvn compile -Dcompile-protobuf
-or
- $ mvn compile -Pcompile-protobuf
-
-After you've done the above, check it and then check in changes (or post a patch
-on a JIRA with your definition file changes and the generated files). Be careful
-to notice new files and files removed and do appropriate git rm/adds.
+Generation of java files from protobuf .proto files included here is done as
+part of the build.

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-endpoint/pom.xml b/hbase-endpoint/pom.xml
index c031d23..5ac1735 100644
--- a/hbase-endpoint/pom.xml
+++ b/hbase-endpoint/pom.xml
@@ -1,6 +1,6 @@
 <?xml version="1.0"?>
 <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
-    <!--
+  <!--
     /**
      * Licensed to the Apache Software Foundation (ASF) under one
      * or more contributor license agreements.  See the NOTICE file
@@ -19,37 +19,34 @@
      * limitations under the License.
      */
     -->
-    <modelVersion>4.0.0</modelVersion>
-    <parent>
-        <artifactId>hbase</artifactId>
-        <groupId>org.apache.hbase</groupId>
-        <version>2.0.0-alpha-2-SNAPSHOT</version>
-        <relativePath>..</relativePath>
-    </parent>
-
-    <artifactId>hbase-endpoint</artifactId>
-    <name>Apache HBase - Coprocessor Endpoint</name>
-    <description>HBase Coprocessor Endpoint implementations</description>
-<!--REMOVE-->
-
-    <properties>
-      <maven.javadoc.skip>true</maven.javadoc.skip>
-    </properties>
-
-    <build>
-      <plugins>
-        <plugin>
-          <groupId>org.apache.maven.plugins</groupId>
-          <artifactId>maven-site-plugin</artifactId>
-          <configuration>
-            <skip>true</skip>
-          </configuration>
-        </plugin>
-        <!-- Make a jar and put the sources in the jar -->
-        <plugin>
-          <groupId>org.apache.maven.plugins</groupId>
-          <artifactId>maven-source-plugin</artifactId>
-        </plugin>
+  <modelVersion>4.0.0</modelVersion>
+  <parent>
+    <artifactId>hbase</artifactId>
+    <groupId>org.apache.hbase</groupId>
+    <version>2.0.0-alpha-2-SNAPSHOT</version>
+    <relativePath>..</relativePath>
+  </parent>
+  <artifactId>hbase-endpoint</artifactId>
+  <name>Apache HBase - Coprocessor Endpoint</name>
+  <description>HBase Coprocessor Endpoint implementations</description>
+  <!--REMOVE-->
+  <properties>
+    <maven.javadoc.skip>true</maven.javadoc.skip>
+  </properties>
+  <build>
+    <plugins>
+      <plugin>
+        <groupId>org.apache.maven.plugins</groupId>
+        <artifactId>maven-site-plugin</artifactId>
+        <configuration>
+          <skip>true</skip>
+        </configuration>
+      </plugin>
+      <!-- Make a jar and put the sources in the jar -->
+      <plugin>
+        <groupId>org.apache.maven.plugins</groupId>
+        <artifactId>maven-source-plugin</artifactId>
+      </plugin>
       <plugin>
         <!--Make it so assembly:single does nothing in here-->
         <artifactId>maven-assembly-plugin</artifactId>
@@ -57,38 +54,56 @@
           <skipAssembly>true</skipAssembly>
         </configuration>
       </plugin>
+      <plugin>
+        <groupId>org.xolstice.maven.plugins</groupId>
+        <artifactId>protobuf-maven-plugin</artifactId>
+        <executions>
+          <execution>
+            <id>compile-protoc</id>
+            <phase>generate-sources</phase>
+            <goals>
+              <goal>compile</goal>
+            </goals>
+             <configuration>
+              <additionalProtoPathElements>
+                <additionalProtoPathElement>${basedir}/../hbase-protocol/src/main/protobuf</additionalProtoPathElement>
+              </additionalProtoPathElements>
+             </configuration>
+          </execution>
+        </executions>
+      </plugin>
+    </plugins>
+    <pluginManagement>
+      <plugins>
+        <!--This plugin's configuration is used to store Eclipse m2e settings only. It has no influence on the Maven build itself.-->
+        <plugin>
+          <groupId>org.eclipse.m2e</groupId>
+          <artifactId>lifecycle-mapping</artifactId>
+          <version>1.0.0</version>
+          <configuration>
+            <lifecycleMappingMetadata>
+              <pluginExecutions>
+                <pluginExecution>
+                  <pluginExecutionFilter>
+                    <groupId>org.apache.hadoop</groupId>
+                    <artifactId>hadoop-maven-plugins</artifactId>
+                    <versionRange>[2.0.5-alpha,)</versionRange>
+                    <goals>
+                      <goal>protoc</goal>
+                    </goals>
+                  </pluginExecutionFilter>
+                  <action>
+                    <ignore/>
+                  </action>
+                </pluginExecution>
+              </pluginExecutions>
+            </lifecycleMappingMetadata>
+          </configuration>
+        </plugin>
       </plugins>
-      <pluginManagement>
-        <plugins>
-          <!--This plugin's configuration is used to store Eclipse m2e settings only. It has no influence on the Maven build itself.-->
-          <plugin>
-            <groupId>org.eclipse.m2e</groupId>
-            <artifactId>lifecycle-mapping</artifactId>
-            <configuration>
-              <lifecycleMappingMetadata>
-                <pluginExecutions>
-                  <pluginExecution>
-                    <pluginExecutionFilter>
-                      <groupId>org.apache.hadoop</groupId>
-                      <artifactId>hadoop-maven-plugins</artifactId>
-                      <versionRange>[2.0.5-alpha,)</versionRange>
-                      <goals>
-                        <goal>protoc</goal>
-                      </goals>
-                    </pluginExecutionFilter>
-                    <action>
-                      <ignore></ignore>
-                    </action>
-                  </pluginExecution>
-                </pluginExecutions>
-              </lifecycleMappingMetadata>
-            </configuration>
-          </plugin>
-         </plugins>
-      </pluginManagement>
-    </build>
-
-    <dependencies>
+    </pluginManagement>
+  </build>
+  <dependencies>
     <dependency>
       <groupId>org.apache.hbase.thirdparty</groupId>
       <artifactId>hbase-shaded-miscellaneous</artifactId>
@@ -131,88 +146,57 @@
       <type>test-jar</type>
       <scope>test</scope>
     </dependency>
-      <dependency>
-        <groupId>org.apache.hbase</groupId>
-        <artifactId>hbase-protocol</artifactId>
-      </dependency>
-      <dependency>
-        <groupId>org.apache.hbase</groupId>
-        <artifactId>hbase-client</artifactId>
-      </dependency>
-      <!--Some of the CPEPs use hbase server-side internals; they shouldn't!
+    <dependency>
+      <groupId>org.apache.hbase</groupId>
+      <artifactId>hbase-protocol</artifactId>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.hbase</groupId>
+      <artifactId>hbase-client</artifactId>
+    </dependency>
+    <!--Some of the CPEPs use hbase server-side internals; they shouldn't!
            -->
-      <dependency>
-        <groupId>org.apache.hbase</groupId>
-        <artifactId>hbase-server</artifactId>
-      </dependency>
-      <dependency>
-        <groupId>org.apache.hbase</groupId>
-        <artifactId>hbase-server</artifactId>
-        <type>test-jar</type>
-        <scope>test</scope>
-      </dependency>
-      <!-- General dependencies -->
-      <dependency>
-        <groupId>commons-logging</groupId>
-        <artifactId>commons-logging</artifactId>
-      </dependency>
-    </dependencies>
-
-    <profiles>
-        <!-- Skip the tests in this module -->
-        <profile>
-            <id>skipRpcTests</id>
-            <activation>
-                <property>
-                    <name>skipRpcTests</name>
-                </property>
-            </activation>
-            <properties>
-                <surefire.skipFirstPart>true</surefire.skipFirstPart>
-                <surefire.skipSecondPart>true</surefire.skipSecondPart>
-            </properties>
-        </profile>
-        <profile>
-          <id>compile-protobuf</id>
-          <activation>
-            <property>
-              <name>compile-protobuf</name>
-            </property>
-          </activation>
-          <build>
-            <plugins>
-              <plugin>
-                <groupId>org.xolstice.maven.plugins</groupId>
-                <artifactId>protobuf-maven-plugin</artifactId>
-                <executions>
-                  <execution>
-                    <id>compile-protoc</id>
-                    <phase>generate-sources</phase>
-                    <goals>
-                      <goal>compile</goal>
-                    </goals>
-                    <configuration>
-                      <additionalProtoPathElements>
-                        <additionalProtoPathElement>${basedir}/../hbase-protocol/src/main/protobuf</additionalProtoPathElement>
-                      </additionalProtoPathElements>
-                    </configuration>
-                  </execution>
-                </executions>
-              </plugin>
-            </plugins>
-          </build>
-        </profile>
+    <dependency>
+      <groupId>org.apache.hbase</groupId>
+      <artifactId>hbase-server</artifactId>
+    </dependency>
+    <dependency>
+      <groupId>org.apache.hbase</groupId>
+      <artifactId>hbase-server</artifactId>
+      <type>test-jar</type>
+      <scope>test</scope>
+    </dependency>
+    <!-- General dependencies -->
+    <dependency>
+      <groupId>commons-logging</groupId>
+      <artifactId>commons-logging</artifactId>
+    </dependency>
+  </dependencies>
+  <profiles>
+    <!-- Skip the tests in this module -->
+    <profile>
+      <id>skipRpcTests</id>
+      <activation>
+        <property>
+          <name>skipRpcTests</name>
+        </property>
+      </activation>
+      <properties>
+        <surefire.skipFirstPart>true</surefire.skipFirstPart>
+        <surefire.skipSecondPart>true</surefire.skipSecondPart>
+      </properties>
+    </profile>
     <!-- Profiles for building against different hadoop versions -->
     <!-- There are a lot of common dependencies used here, should investigate
     if we can combine these profiles somehow -->
-
     <!-- profile for building against Hadoop 2.x.  This is the default.  -->
     <profile>
       <id>hadoop-2.0</id>
       <activation>
         <property>
-            <!--Below formatting for dev-support/generate-hadoopX-poms.sh-->
-            <!--h2--><name>!hadoop.profile</name>
+          <!--Below formatting for dev-support/generate-hadoopX-poms.sh-->
+          <!--h2-->
+          <name>!hadoop.profile</name>
         </property>
       </activation>
       <dependencies>
@@ -315,5 +299,5 @@
         </dependency>
       </dependencies>
     </profile>
-    </profiles>
+  </profiles>
 </project>


[48/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithErrorsProtos.java
----------------------------------------------------------------------
diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithErrorsProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithErrorsProtos.java
deleted file mode 100644
index 64cf82e..0000000
--- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/ColumnAggregationWithErrorsProtos.java
+++ /dev/null
@@ -1,1290 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: ColumnAggregationWithErrorsProtocol.proto
-
-package org.apache.hadoop.hbase.coprocessor.protobuf.generated;
-
-public final class ColumnAggregationWithErrorsProtos {
-  private ColumnAggregationWithErrorsProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface ColumnAggregationWithErrorsSumRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required bytes family = 1;
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    boolean hasFamily();
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    com.google.protobuf.ByteString getFamily();
-
-    // optional bytes qualifier = 2;
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    boolean hasQualifier();
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    com.google.protobuf.ByteString getQualifier();
-  }
-  /**
-   * Protobuf type {@code ColumnAggregationWithErrorsSumRequest}
-   *
-   * <pre>
-   * use unique names for messages in ColumnAggregationXXX.protos due to a bug in
-   * protoc or hadoop's protoc compiler.
-   * </pre>
-   */
-  public static final class ColumnAggregationWithErrorsSumRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements ColumnAggregationWithErrorsSumRequestOrBuilder {
-    // Use ColumnAggregationWithErrorsSumRequest.newBuilder() to construct.
-    private ColumnAggregationWithErrorsSumRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private ColumnAggregationWithErrorsSumRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final ColumnAggregationWithErrorsSumRequest defaultInstance;
-    public static ColumnAggregationWithErrorsSumRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public ColumnAggregationWithErrorsSumRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private ColumnAggregationWithErrorsSumRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              family_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              bitField0_ |= 0x00000002;
-              qualifier_ = input.readBytes();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<ColumnAggregationWithErrorsSumRequest> PARSER =
-        new com.google.protobuf.AbstractParser<ColumnAggregationWithErrorsSumRequest>() {
-      public ColumnAggregationWithErrorsSumRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new ColumnAggregationWithErrorsSumRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<ColumnAggregationWithErrorsSumRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required bytes family = 1;
-    public static final int FAMILY_FIELD_NUMBER = 1;
-    private com.google.protobuf.ByteString family_;
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    public boolean hasFamily() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required bytes family = 1;</code>
-     */
-    public com.google.protobuf.ByteString getFamily() {
-      return family_;
-    }
-
-    // optional bytes qualifier = 2;
-    public static final int QUALIFIER_FIELD_NUMBER = 2;
-    private com.google.protobuf.ByteString qualifier_;
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    public boolean hasQualifier() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>optional bytes qualifier = 2;</code>
-     */
-    public com.google.protobuf.ByteString getQualifier() {
-      return qualifier_;
-    }
-
-    private void initFields() {
-      family_ = com.google.protobuf.ByteString.EMPTY;
-      qualifier_ = com.google.protobuf.ByteString.EMPTY;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasFamily()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, family_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeBytes(2, qualifier_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, family_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(2, qualifier_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest) obj;
-
-      boolean result = true;
-      result = result && (hasFamily() == other.hasFamily());
-      if (hasFamily()) {
-        result = result && getFamily()
-            .equals(other.getFamily());
-      }
-      result = result && (hasQualifier() == other.hasQualifier());
-      if (hasQualifier()) {
-        result = result && getQualifier()
-            .equals(other.getQualifier());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasFamily()) {
-        hash = (37 * hash) + FAMILY_FIELD_NUMBER;
-        hash = (53 * hash) + getFamily().hashCode();
-      }
-      if (hasQualifier()) {
-        hash = (37 * hash) + QUALIFIER_FIELD_NUMBER;
-        hash = (53 * hash) + getQualifier().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code ColumnAggregationWithErrorsSumRequest}
-     *
-     * <pre>
-     * use unique names for messages in ColumnAggregationXXX.protos due to a bug in
-     * protoc or hadoop's protoc compiler.
-     * </pre>
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        family_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        qualifier_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.family_ = family_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.qualifier_ = qualifier_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.getDefaultInstance()) return this;
-        if (other.hasFamily()) {
-          setFamily(other.getFamily());
-        }
-        if (other.hasQualifier()) {
-          setQualifier(other.getQualifier());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasFamily()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required bytes family = 1;
-      private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public boolean hasFamily() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public com.google.protobuf.ByteString getFamily() {
-        return family_;
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public Builder setFamily(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        family_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required bytes family = 1;</code>
-       */
-      public Builder clearFamily() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        family_ = getDefaultInstance().getFamily();
-        onChanged();
-        return this;
-      }
-
-      // optional bytes qualifier = 2;
-      private com.google.protobuf.ByteString qualifier_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public boolean hasQualifier() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public com.google.protobuf.ByteString getQualifier() {
-        return qualifier_;
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public Builder setQualifier(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000002;
-        qualifier_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>optional bytes qualifier = 2;</code>
-       */
-      public Builder clearQualifier() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        qualifier_ = getDefaultInstance().getQualifier();
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:ColumnAggregationWithErrorsSumRequest)
-    }
-
-    static {
-      defaultInstance = new ColumnAggregationWithErrorsSumRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:ColumnAggregationWithErrorsSumRequest)
-  }
-
-  public interface ColumnAggregationWithErrorsSumResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required int64 sum = 1;
-    /**
-     * <code>required int64 sum = 1;</code>
-     */
-    boolean hasSum();
-    /**
-     * <code>required int64 sum = 1;</code>
-     */
-    long getSum();
-  }
-  /**
-   * Protobuf type {@code ColumnAggregationWithErrorsSumResponse}
-   */
-  public static final class ColumnAggregationWithErrorsSumResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements ColumnAggregationWithErrorsSumResponseOrBuilder {
-    // Use ColumnAggregationWithErrorsSumResponse.newBuilder() to construct.
-    private ColumnAggregationWithErrorsSumResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private ColumnAggregationWithErrorsSumResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final ColumnAggregationWithErrorsSumResponse defaultInstance;
-    public static ColumnAggregationWithErrorsSumResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public ColumnAggregationWithErrorsSumResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private ColumnAggregationWithErrorsSumResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              sum_ = input.readInt64();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<ColumnAggregationWithErrorsSumResponse> PARSER =
-        new com.google.protobuf.AbstractParser<ColumnAggregationWithErrorsSumResponse>() {
-      public ColumnAggregationWithErrorsSumResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new ColumnAggregationWithErrorsSumResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<ColumnAggregationWithErrorsSumResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required int64 sum = 1;
-    public static final int SUM_FIELD_NUMBER = 1;
-    private long sum_;
-    /**
-     * <code>required int64 sum = 1;</code>
-     */
-    public boolean hasSum() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required int64 sum = 1;</code>
-     */
-    public long getSum() {
-      return sum_;
-    }
-
-    private void initFields() {
-      sum_ = 0L;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasSum()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeInt64(1, sum_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt64Size(1, sum_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse) obj;
-
-      boolean result = true;
-      result = result && (hasSum() == other.hasSum());
-      if (hasSum()) {
-        result = result && (getSum()
-            == other.getSum());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasSum()) {
-        hash = (37 * hash) + SUM_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getSum());
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code ColumnAggregationWithErrorsSumResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        sum_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.internal_static_ColumnAggregationWithErrorsSumResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.sum_ = sum_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.getDefaultInstance()) return this;
-        if (other.hasSum()) {
-          setSum(other.getSum());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasSum()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required int64 sum = 1;
-      private long sum_ ;
-      /**
-       * <code>required int64 sum = 1;</code>
-       */
-      public boolean hasSum() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required int64 sum = 1;</code>
-       */
-      public long getSum() {
-        return sum_;
-      }
-      /**
-       * <code>required int64 sum = 1;</code>
-       */
-      public Builder setSum(long value) {
-        bitField0_ |= 0x00000001;
-        sum_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required int64 sum = 1;</code>
-       */
-      public Builder clearSum() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        sum_ = 0L;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:ColumnAggregationWithErrorsSumResponse)
-    }
-
-    static {
-      defaultInstance = new ColumnAggregationWithErrorsSumResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:ColumnAggregationWithErrorsSumResponse)
-  }
-
-  /**
-   * Protobuf service {@code ColumnAggregationServiceWithErrors}
-   */
-  public static abstract class ColumnAggregationServiceWithErrors
-      implements com.google.protobuf.Service {
-    protected ColumnAggregationServiceWithErrors() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc sum(.ColumnAggregationWithErrorsSumRequest) returns (.ColumnAggregationWithErrorsSumResponse);</code>
-       */
-      public abstract void sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse> done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new ColumnAggregationServiceWithErrors() {
-        @java.lang.Override
-        public  void sum(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse> done) {
-          impl.sum(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc sum(.ColumnAggregationWithErrorsSumRequest) returns (.ColumnAggregationWithErrorsSumResponse);</code>
-     */
-    public abstract void sum(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse> done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.sum(controller, (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationServiceWithErrors implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.class,
-            org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest request)
-          throws com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse sum(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.protobuf.generated.ColumnAggregationWithErrorsProtos.ColumnAggregationWithErrorsSumResponse.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:ColumnAggregationServiceWithErrors)
-  }
-
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_ColumnAggregationWithErrorsSumRequest_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_ColumnAggregationWithErrorsSumRequest_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_ColumnAggregationWithErrorsSumResponse_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_ColumnAggregationWithErrorsSumResponse_fieldAccessorTable;
-
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n)ColumnAggregationWithErrorsProtocol.pr" +
-      "oto\"J\n%ColumnAggregationWithErrorsSumReq" +
-      "uest\022\016\n\006family\030\001 \002(\014\022\021\n\tqualifier\030\002 \001(\014\"" +
-      "5\n&ColumnAggregationWithErrorsSumRespons" +
-      "e\022\013\n\003sum\030\001 \002(\0032|\n\"ColumnAggregationServi" +
-      "ceWithErrors\022V\n\003sum\022&.ColumnAggregationW" +
-      "ithErrorsSumRequest\032\'.ColumnAggregationW" +
-      "ithErrorsSumResponseBa\n6org.apache.hadoo" +
-      "p.hbase.coprocessor.protobuf.generatedB!" +
-      "ColumnAggregationWithErrorsProtos\210\001\001\240\001\001"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_ColumnAggregationWithErrorsSumRequest_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_ColumnAggregationWithErrorsSumRequest_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_ColumnAggregationWithErrorsSumRequest_descriptor,
-              new java.lang.String[] { "Family", "Qualifier", });
-          internal_static_ColumnAggregationWithErrorsSumResponse_descriptor =
-            getDescriptor().getMessageTypes().get(1);
-          internal_static_ColumnAggregationWithErrorsSumResponse_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_ColumnAggregationWithErrorsSumResponse_descriptor,
-              new java.lang.String[] { "Sum", });
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[15/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StringValue.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StringValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StringValue.java
deleted file mode 100644
index f3e2ef6..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StringValue.java
+++ /dev/null
@@ -1,532 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `string`.
- * The JSON representation for `StringValue` is JSON string.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.StringValue}
- */
-public  final class StringValue extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.StringValue)
-    StringValueOrBuilder {
-  // Use StringValue.newBuilder() to construct.
-  private StringValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private StringValue() {
-    value_ = "";
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private StringValue(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            value_ = s;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_StringValue_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_StringValue_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private volatile java.lang.Object value_;
-  /**
-   * <pre>
-   * The string value.
-   * </pre>
-   *
-   * <code>string value = 1;</code>
-   */
-  public java.lang.String getValue() {
-    java.lang.Object ref = value_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      value_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The string value.
-   * </pre>
-   *
-   * <code>string value = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getValueBytes() {
-    java.lang.Object ref = value_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      value_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getValueBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getValueBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue) obj;
-
-    boolean result = true;
-    result = result && getValue()
-        .equals(other.getValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + getValue().hashCode();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `string`.
-   * The JSON representation for `StringValue` is JSON string.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.StringValue}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.StringValue)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_StringValue_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_StringValue_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = "";
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_StringValue_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue.getDefaultInstance()) return this;
-      if (!other.getValue().isEmpty()) {
-        value_ = other.value_;
-        onChanged();
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private java.lang.Object value_ = "";
-    /**
-     * <pre>
-     * The string value.
-     * </pre>
-     *
-     * <code>string value = 1;</code>
-     */
-    public java.lang.String getValue() {
-      java.lang.Object ref = value_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        value_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The string value.
-     * </pre>
-     *
-     * <code>string value = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getValueBytes() {
-      java.lang.Object ref = value_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        value_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The string value.
-     * </pre>
-     *
-     * <code>string value = 1;</code>
-     */
-    public Builder setValue(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The string value.
-     * </pre>
-     *
-     * <code>string value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = getDefaultInstance().getValue();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The string value.
-     * </pre>
-     *
-     * <code>string value = 1;</code>
-     */
-    public Builder setValueBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.StringValue)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.StringValue)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StringValue>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<StringValue>() {
-    public StringValue parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new StringValue(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StringValue> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<StringValue> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.StringValue getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StringValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StringValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StringValueOrBuilder.java
deleted file mode 100644
index 7ac0ee8..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StringValueOrBuilder.java
+++ /dev/null
@@ -1,27 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface StringValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.StringValue)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The string value.
-   * </pre>
-   *
-   * <code>string value = 1;</code>
-   */
-  java.lang.String getValue();
-  /**
-   * <pre>
-   * The string value.
-   * </pre>
-   *
-   * <code>string value = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getValueBytes();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Struct.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Struct.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Struct.java
deleted file mode 100644
index 2607900..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Struct.java
+++ /dev/null
@@ -1,705 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/struct.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * `Struct` represents a structured data value, consisting of fields
- * which map to dynamically typed values. In some languages, `Struct`
- * might be supported by a native representation. For example, in
- * scripting languages like JS a struct is represented as an
- * object. The details of that representation are described together
- * with the proto support for the language.
- * The JSON representation for `Struct` is JSON object.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Struct}
- */
-public  final class Struct extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Struct)
-    StructOrBuilder {
-  // Use Struct.newBuilder() to construct.
-  private Struct(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Struct() {
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Struct(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-              fields_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField.newMapField(
-                  FieldsDefaultEntryHolder.defaultEntry);
-              mutable_bitField0_ |= 0x00000001;
-            }
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.MapEntry<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>
-            fields__ = input.readMessage(
-                FieldsDefaultEntryHolder.defaultEntry.getParserForType(), extensionRegistry);
-            fields_.getMutableMap().put(
-                fields__.getKey(), fields__.getValue());
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Struct_descriptor;
-  }
-
-  @SuppressWarnings({"rawtypes"})
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField internalGetMapField(
-      int number) {
-    switch (number) {
-      case 1:
-        return internalGetFields();
-      default:
-        throw new RuntimeException(
-            "Invalid map field number: " + number);
-    }
-  }
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Struct_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.Builder.class);
-  }
-
-  public static final int FIELDS_FIELD_NUMBER = 1;
-  private static final class FieldsDefaultEntryHolder {
-    static final org.apache.hadoop.hbase.shaded.com.google.protobuf.MapEntry<
-        java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> defaultEntry =
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.MapEntry
-            .<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>newDefaultInstance(
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Struct_FieldsEntry_descriptor, 
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.FieldType.STRING,
-                "",
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.WireFormat.FieldType.MESSAGE,
-                org.apache.hadoop.hbase.shaded.com.google.protobuf.Value.getDefaultInstance());
-  }
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField<
-      java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> fields_;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>
-  internalGetFields() {
-    if (fields_ == null) {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField.emptyMapField(
-          FieldsDefaultEntryHolder.defaultEntry);
-    }
-    return fields_;
-  }
-
-  public int getFieldsCount() {
-    return internalGetFields().getMap().size();
-  }
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-
-  public boolean containsFields(
-      java.lang.String key) {
-    if (key == null) { throw new java.lang.NullPointerException(); }
-    return internalGetFields().getMap().containsKey(key);
-  }
-  /**
-   * Use {@link #getFieldsMap()} instead.
-   */
-  @java.lang.Deprecated
-  public java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> getFields() {
-    return getFieldsMap();
-  }
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-
-  public java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> getFieldsMap() {
-    return internalGetFields().getMap();
-  }
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getFieldsOrDefault(
-      java.lang.String key,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Value defaultValue) {
-    if (key == null) { throw new java.lang.NullPointerException(); }
-    java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> map =
-        internalGetFields().getMap();
-    return map.containsKey(key) ? map.get(key) : defaultValue;
-  }
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getFieldsOrThrow(
-      java.lang.String key) {
-    if (key == null) { throw new java.lang.NullPointerException(); }
-    java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> map =
-        internalGetFields().getMap();
-    if (!map.containsKey(key)) {
-      throw new java.lang.IllegalArgumentException();
-    }
-    return map.get(key);
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-      .serializeStringMapTo(
-        output,
-        internalGetFields(),
-        FieldsDefaultEntryHolder.defaultEntry,
-        1);
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    for (java.util.Map.Entry<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> entry
-         : internalGetFields().getMap().entrySet()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.MapEntry<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>
-      fields__ = FieldsDefaultEntryHolder.defaultEntry.newBuilderForType()
-          .setKey(entry.getKey())
-          .setValue(entry.getValue())
-          .build();
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, fields__);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) obj;
-
-    boolean result = true;
-    result = result && internalGetFields().equals(
-        other.internalGetFields());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    if (!internalGetFields().getMap().isEmpty()) {
-      hash = (37 * hash) + FIELDS_FIELD_NUMBER;
-      hash = (53 * hash) + internalGetFields().hashCode();
-    }
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * `Struct` represents a structured data value, consisting of fields
-   * which map to dynamically typed values. In some languages, `Struct`
-   * might be supported by a native representation. For example, in
-   * scripting languages like JS a struct is represented as an
-   * object. The details of that representation are described together
-   * with the proto support for the language.
-   * The JSON representation for `Struct` is JSON object.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Struct}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Struct)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.StructOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Struct_descriptor;
-    }
-
-    @SuppressWarnings({"rawtypes"})
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField internalGetMapField(
-        int number) {
-      switch (number) {
-        case 1:
-          return internalGetFields();
-        default:
-          throw new RuntimeException(
-              "Invalid map field number: " + number);
-      }
-    }
-    @SuppressWarnings({"rawtypes"})
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField internalGetMutableMapField(
-        int number) {
-      switch (number) {
-        case 1:
-          return internalGetMutableFields();
-        default:
-          throw new RuntimeException(
-              "Invalid map field number: " + number);
-      }
-    }
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Struct_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      internalGetMutableFields().clear();
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.StructProto.internal_static_google_protobuf_Struct_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct(this);
-      int from_bitField0_ = bitField0_;
-      result.fields_ = internalGetFields();
-      result.fields_.makeImmutable();
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct.getDefaultInstance()) return this;
-      internalGetMutableFields().mergeFrom(
-          other.internalGetFields());
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-    private int bitField0_;
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField<
-        java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> fields_;
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>
-    internalGetFields() {
-      if (fields_ == null) {
-        return org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField.emptyMapField(
-            FieldsDefaultEntryHolder.defaultEntry);
-      }
-      return fields_;
-    }
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>
-    internalGetMutableFields() {
-      onChanged();;
-      if (fields_ == null) {
-        fields_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.MapField.newMapField(
-            FieldsDefaultEntryHolder.defaultEntry);
-      }
-      if (!fields_.isMutable()) {
-        fields_ = fields_.copy();
-      }
-      return fields_;
-    }
-
-    public int getFieldsCount() {
-      return internalGetFields().getMap().size();
-    }
-    /**
-     * <pre>
-     * Unordered map of dynamically typed values.
-     * </pre>
-     *
-     * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-     */
-
-    public boolean containsFields(
-        java.lang.String key) {
-      if (key == null) { throw new java.lang.NullPointerException(); }
-      return internalGetFields().getMap().containsKey(key);
-    }
-    /**
-     * Use {@link #getFieldsMap()} instead.
-     */
-    @java.lang.Deprecated
-    public java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> getFields() {
-      return getFieldsMap();
-    }
-    /**
-     * <pre>
-     * Unordered map of dynamically typed values.
-     * </pre>
-     *
-     * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-     */
-
-    public java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> getFieldsMap() {
-      return internalGetFields().getMap();
-    }
-    /**
-     * <pre>
-     * Unordered map of dynamically typed values.
-     * </pre>
-     *
-     * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-     */
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getFieldsOrDefault(
-        java.lang.String key,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Value defaultValue) {
-      if (key == null) { throw new java.lang.NullPointerException(); }
-      java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> map =
-          internalGetFields().getMap();
-      return map.containsKey(key) ? map.get(key) : defaultValue;
-    }
-    /**
-     * <pre>
-     * Unordered map of dynamically typed values.
-     * </pre>
-     *
-     * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-     */
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getFieldsOrThrow(
-        java.lang.String key) {
-      if (key == null) { throw new java.lang.NullPointerException(); }
-      java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> map =
-          internalGetFields().getMap();
-      if (!map.containsKey(key)) {
-        throw new java.lang.IllegalArgumentException();
-      }
-      return map.get(key);
-    }
-
-    public Builder clearFields() {
-      internalGetMutableFields().getMutableMap()
-          .clear();
-      return this;
-    }
-    /**
-     * <pre>
-     * Unordered map of dynamically typed values.
-     * </pre>
-     *
-     * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-     */
-
-    public Builder removeFields(
-        java.lang.String key) {
-      if (key == null) { throw new java.lang.NullPointerException(); }
-      internalGetMutableFields().getMutableMap()
-          .remove(key);
-      return this;
-    }
-    /**
-     * Use alternate mutation accessors instead.
-     */
-    @java.lang.Deprecated
-    public java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>
-    getMutableFields() {
-      return internalGetMutableFields().getMutableMap();
-    }
-    /**
-     * <pre>
-     * Unordered map of dynamically typed values.
-     * </pre>
-     *
-     * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-     */
-    public Builder putFields(
-        java.lang.String key,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Value value) {
-      if (key == null) { throw new java.lang.NullPointerException(); }
-      if (value == null) { throw new java.lang.NullPointerException(); }
-      internalGetMutableFields().getMutableMap()
-          .put(key, value);
-      return this;
-    }
-    /**
-     * <pre>
-     * Unordered map of dynamically typed values.
-     * </pre>
-     *
-     * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-     */
-
-    public Builder putAllFields(
-        java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value> values) {
-      internalGetMutableFields().getMutableMap()
-          .putAll(values);
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Struct)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Struct)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Struct>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Struct>() {
-    public Struct parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Struct(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Struct> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Struct> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Struct getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StructOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StructOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StructOrBuilder.java
deleted file mode 100644
index 229f540..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StructOrBuilder.java
+++ /dev/null
@@ -1,63 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/struct.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface StructOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Struct)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-  int getFieldsCount();
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-  boolean containsFields(
-      java.lang.String key);
-  /**
-   * Use {@link #getFieldsMap()} instead.
-   */
-  @java.lang.Deprecated
-  java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>
-  getFields();
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-  java.util.Map<java.lang.String, org.apache.hadoop.hbase.shaded.com.google.protobuf.Value>
-  getFieldsMap();
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getFieldsOrDefault(
-      java.lang.String key,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Value defaultValue);
-  /**
-   * <pre>
-   * Unordered map of dynamically typed values.
-   * </pre>
-   *
-   * <code>map&lt;string, .google.protobuf.Value&gt; fields = 1;</code>
-   */
-
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.Value getFieldsOrThrow(
-      java.lang.String key);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StructProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StructProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StructProto.java
deleted file mode 100644
index 4d5cd4b..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/StructProto.java
+++ /dev/null
@@ -1,103 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/struct.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class StructProto {
-  private StructProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Struct_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Struct_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Struct_FieldsEntry_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Struct_FieldsEntry_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Value_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Value_fieldAccessorTable;
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_ListValue_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_ListValue_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\034google/protobuf/struct.proto\022\017google.p" +
-      "rotobuf\"\204\001\n\006Struct\0223\n\006fields\030\001 \003(\0132#.goo" +
-      "gle.protobuf.Struct.FieldsEntry\032E\n\013Field" +
-      "sEntry\022\013\n\003key\030\001 \001(\t\022%\n\005value\030\002 \001(\0132\026.goo" +
-      "gle.protobuf.Value:\0028\001\"\352\001\n\005Value\0220\n\nnull" +
-      "_value\030\001 \001(\0162\032.google.protobuf.NullValue" +
-      "H\000\022\026\n\014number_value\030\002 \001(\001H\000\022\026\n\014string_val" +
-      "ue\030\003 \001(\tH\000\022\024\n\nbool_value\030\004 \001(\010H\000\022/\n\014stru" +
-      "ct_value\030\005 \001(\0132\027.google.protobuf.StructH" +
-      "\000\0220\n\nlist_value\030\006 \001(\0132\032.google.protobuf.",
-      "ListValueH\000B\006\n\004kind\"3\n\tListValue\022&\n\006valu" +
-      "es\030\001 \003(\0132\026.google.protobuf.Value*\033\n\tNull" +
-      "Value\022\016\n\nNULL_VALUE\020\000B\201\001\n\023com.google.pro" +
-      "tobufB\013StructProtoP\001Z1github.com/golang/" +
-      "protobuf/ptypes/struct;structpb\370\001\001\242\002\003GPB" +
-      "\252\002\036Google.Protobuf.WellKnownTypesb\006proto" +
-      "3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_google_protobuf_Struct_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_Struct_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Struct_descriptor,
-        new java.lang.String[] { "Fields", });
-    internal_static_google_protobuf_Struct_FieldsEntry_descriptor =
-      internal_static_google_protobuf_Struct_descriptor.getNestedTypes().get(0);
-    internal_static_google_protobuf_Struct_FieldsEntry_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Struct_FieldsEntry_descriptor,
-        new java.lang.String[] { "Key", "Value", });
-    internal_static_google_protobuf_Value_descriptor =
-      getDescriptor().getMessageTypes().get(1);
-    internal_static_google_protobuf_Value_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Value_descriptor,
-        new java.lang.String[] { "NullValue", "NumberValue", "StringValue", "BoolValue", "StructValue", "ListValue", "Kind", });
-    internal_static_google_protobuf_ListValue_descriptor =
-      getDescriptor().getMessageTypes().get(2);
-    internal_static_google_protobuf_ListValue_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_ListValue_descriptor,
-        new java.lang.String[] { "Values", });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Syntax.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Syntax.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Syntax.java
deleted file mode 100644
index f03d70b..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Syntax.java
+++ /dev/null
@@ -1,123 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/type.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * The syntax in which a protocol buffer element is defined.
- * </pre>
- *
- * Protobuf enum {@code google.protobuf.Syntax}
- */
-public enum Syntax
-    implements org.apache.hadoop.hbase.shaded.com.google.protobuf.ProtocolMessageEnum {
-  /**
-   * <pre>
-   * Syntax `proto2`.
-   * </pre>
-   *
-   * <code>SYNTAX_PROTO2 = 0;</code>
-   */
-  SYNTAX_PROTO2(0),
-  /**
-   * <pre>
-   * Syntax `proto3`.
-   * </pre>
-   *
-   * <code>SYNTAX_PROTO3 = 1;</code>
-   */
-  SYNTAX_PROTO3(1),
-  UNRECOGNIZED(-1),
-  ;
-
-  /**
-   * <pre>
-   * Syntax `proto2`.
-   * </pre>
-   *
-   * <code>SYNTAX_PROTO2 = 0;</code>
-   */
-  public static final int SYNTAX_PROTO2_VALUE = 0;
-  /**
-   * <pre>
-   * Syntax `proto3`.
-   * </pre>
-   *
-   * <code>SYNTAX_PROTO3 = 1;</code>
-   */
-  public static final int SYNTAX_PROTO3_VALUE = 1;
-
-
-  public final int getNumber() {
-    if (this == UNRECOGNIZED) {
-      throw new java.lang.IllegalArgumentException(
-          "Can't get the number of an unknown enum value.");
-    }
-    return value;
-  }
-
-  /**
-   * @deprecated Use {@link #forNumber(int)} instead.
-   */
-  @java.lang.Deprecated
-  public static Syntax valueOf(int value) {
-    return forNumber(value);
-  }
-
-  public static Syntax forNumber(int value) {
-    switch (value) {
-      case 0: return SYNTAX_PROTO2;
-      case 1: return SYNTAX_PROTO3;
-      default: return null;
-    }
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Syntax>
-      internalGetValueMap() {
-    return internalValueMap;
-  }
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<
-      Syntax> internalValueMap =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.EnumLiteMap<Syntax>() {
-          public Syntax findValueByNumber(int number) {
-            return Syntax.forNumber(number);
-          }
-        };
-
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor
-      getValueDescriptor() {
-    return getDescriptor().getValues().get(ordinal());
-  }
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-      getDescriptorForType() {
-    return getDescriptor();
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumDescriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.getDescriptor().getEnumTypes().get(0);
-  }
-
-  private static final Syntax[] VALUES = values();
-
-  public static Syntax valueOf(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
-    if (desc.getType() != getDescriptor()) {
-      throw new java.lang.IllegalArgumentException(
-        "EnumValueDescriptor is not for this type.");
-    }
-    if (desc.getIndex() == -1) {
-      return UNRECOGNIZED;
-    }
-    return VALUES[desc.getIndex()];
-  }
-
-  private final int value;
-
-  private Syntax(int value) {
-    this.value = value;
-  }
-
-  // @@protoc_insertion_point(enum_scope:google.protobuf.Syntax)
-}
-


[33/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Descriptors.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Descriptors.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Descriptors.java
deleted file mode 100644
index 62ccd19..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Descriptors.java
+++ /dev/null
@@ -1,2547 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.DescriptorProtos.*;
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.Syntax;
-
-import java.lang.ref.WeakReference;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.HashSet;
-import java.util.List;
-import java.util.Map;
-import java.util.Set;
-import java.util.WeakHashMap;
-import java.util.logging.Logger;
-
-/**
- * Contains a collection of classes which describe protocol message types.
- *
- * Every message type has a {@link Descriptor}, which lists all
- * its fields and other information about a type.  You can get a message
- * type's descriptor by calling {@code MessageType.getDescriptor()}, or
- * (given a message object of the type) {@code message.getDescriptorForType()}.
- * Furthermore, each message is associated with a {@link FileDescriptor} for
- * a relevant {@code .proto} file. You can obtain it by calling
- * {@code Descriptor.getFile()}. A {@link FileDescriptor} contains descriptors
- * for all the messages defined in that file, and file descriptors for all the
- * imported {@code .proto} files.
- *
- * Descriptors are built from DescriptorProtos, as defined in
- * {@code google/protobuf/descriptor.proto}.
- *
- * @author kenton@google.com Kenton Varda
- */
-public final class Descriptors {
-  private static final Logger logger =
-      Logger.getLogger(Descriptors.class.getName());
-  /**
-   * Describes a {@code .proto} file, including everything defined within.
-   * That includes, in particular, descriptors for all the messages and
-   * file descriptors for all other imported {@code .proto} files
-   * (dependencies).
-   */
-  public static final class FileDescriptor extends GenericDescriptor {
-    /** Convert the descriptor to its protocol message representation. */
-    @Override
-    public FileDescriptorProto toProto() {
-      return proto;
-    }
-
-    /** Get the file name. */
-    @Override
-    public String getName() {
-      return proto.getName();
-    }
-
-    /** Returns this object. */
-    @Override
-    public FileDescriptor getFile() {
-      return this;
-    }
-
-    /** Returns the same as getName(). */
-    @Override
-    public String getFullName() {
-      return proto.getName();
-    }
-
-    /**
-     * Get the proto package name.  This is the package name given by the
-     * {@code package} statement in the {@code .proto} file, which differs
-     * from the Java package.
-     */
-    public String getPackage() { return proto.getPackage(); }
-
-    /** Get the {@code FileOptions}, defined in {@code descriptor.proto}. */
-    public FileOptions getOptions() { return proto.getOptions(); }
-
-    /** Get a list of top-level message types declared in this file. */
-    public List<Descriptor> getMessageTypes() {
-      return Collections.unmodifiableList(Arrays.asList(messageTypes));
-    }
-
-    /** Get a list of top-level enum types declared in this file. */
-    public List<EnumDescriptor> getEnumTypes() {
-      return Collections.unmodifiableList(Arrays.asList(enumTypes));
-    }
-
-    /** Get a list of top-level services declared in this file. */
-    public List<ServiceDescriptor> getServices() {
-      return Collections.unmodifiableList(Arrays.asList(services));
-    }
-
-    /** Get a list of top-level extensions declared in this file. */
-    public List<FieldDescriptor> getExtensions() {
-      return Collections.unmodifiableList(Arrays.asList(extensions));
-    }
-
-    /** Get a list of this file's dependencies (imports). */
-    public List<FileDescriptor> getDependencies() {
-      return Collections.unmodifiableList(Arrays.asList(dependencies));
-    }
-
-    /** Get a list of this file's public dependencies (public imports). */
-    public List<FileDescriptor> getPublicDependencies() {
-      return Collections.unmodifiableList(Arrays.asList(publicDependencies));
-    }
-
-    /** The syntax of the .proto file. */
-    public enum Syntax {
-      UNKNOWN("unknown"),
-      PROTO2("proto2"),
-      PROTO3("proto3");
-
-      Syntax(String name) {
-        this.name = name;
-      }
-      private final String name;
-    }
-
-    /** Get the syntax of the .proto file. */
-    public Syntax getSyntax() {
-      if (Syntax.PROTO3.name.equals(proto.getSyntax())) {
-        return Syntax.PROTO3;
-      }
-      return Syntax.PROTO2;
-    }
-
-    /**
-     * Find a message type in the file by name.  Does not find nested types.
-     *
-     * @param name The unqualified type name to look for.
-     * @return The message type's descriptor, or {@code null} if not found.
-     */
-    public Descriptor findMessageTypeByName(String name) {
-      // Don't allow looking up nested types.  This will make optimization
-      // easier later.
-      if (name.indexOf('.') != -1) {
-        return null;
-      }
-      if (getPackage().length() > 0) {
-        name = getPackage() + '.' + name;
-      }
-      final GenericDescriptor result = pool.findSymbol(name);
-      if (result != null && result instanceof Descriptor &&
-          result.getFile() == this) {
-        return (Descriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    /**
-     * Find an enum type in the file by name.  Does not find nested types.
-     *
-     * @param name The unqualified type name to look for.
-     * @return The enum type's descriptor, or {@code null} if not found.
-     */
-    public EnumDescriptor findEnumTypeByName(String name) {
-      // Don't allow looking up nested types.  This will make optimization
-      // easier later.
-      if (name.indexOf('.') != -1) {
-        return null;
-      }
-      if (getPackage().length() > 0) {
-        name = getPackage() + '.' + name;
-      }
-      final GenericDescriptor result = pool.findSymbol(name);
-      if (result != null && result instanceof EnumDescriptor &&
-          result.getFile() == this) {
-        return (EnumDescriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    /**
-     * Find a service type in the file by name.
-     *
-     * @param name The unqualified type name to look for.
-     * @return The service type's descriptor, or {@code null} if not found.
-     */
-    public ServiceDescriptor findServiceByName(String name) {
-      // Don't allow looking up nested types.  This will make optimization
-      // easier later.
-      if (name.indexOf('.') != -1) {
-        return null;
-      }
-      if (getPackage().length() > 0) {
-        name = getPackage() + '.' + name;
-      }
-      final GenericDescriptor result = pool.findSymbol(name);
-      if (result != null && result instanceof ServiceDescriptor &&
-          result.getFile() == this) {
-        return (ServiceDescriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    /**
-     * Find an extension in the file by name.  Does not find extensions nested
-     * inside message types.
-     *
-     * @param name The unqualified extension name to look for.
-     * @return The extension's descriptor, or {@code null} if not found.
-     */
-    public FieldDescriptor findExtensionByName(String name) {
-      if (name.indexOf('.') != -1) {
-        return null;
-      }
-      if (getPackage().length() > 0) {
-        name = getPackage() + '.' + name;
-      }
-      final GenericDescriptor result = pool.findSymbol(name);
-      if (result != null && result instanceof FieldDescriptor &&
-          result.getFile() == this) {
-        return (FieldDescriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    /**
-     * Construct a {@code FileDescriptor}.
-     *
-     * @param proto The protocol message form of the FileDescriptor.
-     * @param dependencies {@code FileDescriptor}s corresponding to all of
-     *                     the file's dependencies.
-     * @throws DescriptorValidationException {@code proto} is not a valid
-     *           descriptor.  This can occur for a number of reasons, e.g.
-     *           because a field has an undefined type or because two messages
-     *           were defined with the same name.
-     */
-    public static FileDescriptor buildFrom(final FileDescriptorProto proto,
-                                           final FileDescriptor[] dependencies)
-                                    throws DescriptorValidationException {
-      return buildFrom(proto, dependencies, false);
-    }
-
-
-    /**
-     * Construct a {@code FileDescriptor}.
-     *
-     * @param proto The protocol message form of the FileDescriptor.
-     * @param dependencies {@code FileDescriptor}s corresponding to all of
-     *                     the file's dependencies.
-     * @param allowUnknownDependencies If true, non-exist dependenncies will be
-     *           ignored and undefined message types will be replaced with a
-     *           placeholder type.
-     * @throws DescriptorValidationException {@code proto} is not a valid
-     *           descriptor.  This can occur for a number of reasons, e.g.
-     *           because a field has an undefined type or because two messages
-     *           were defined with the same name.
-     */
-    public static FileDescriptor buildFrom(
-        final FileDescriptorProto proto, final FileDescriptor[] dependencies,
-        final boolean allowUnknownDependencies)
-        throws DescriptorValidationException {
-      // Building descriptors involves two steps:  translating and linking.
-      // In the translation step (implemented by FileDescriptor's
-      // constructor), we build an object tree mirroring the
-      // FileDescriptorProto's tree and put all of the descriptors into the
-      // DescriptorPool's lookup tables.  In the linking step, we look up all
-      // type references in the DescriptorPool, so that, for example, a
-      // FieldDescriptor for an embedded message contains a pointer directly
-      // to the Descriptor for that message's type.  We also detect undefined
-      // types in the linking step.
-      final DescriptorPool pool = new DescriptorPool(
-          dependencies, allowUnknownDependencies);
-      final FileDescriptor result = new FileDescriptor(
-          proto, dependencies, pool, allowUnknownDependencies);
-      result.crossLink();
-      return result;
-    }
-
-    /**
-     * This method is to be called by generated code only.  It is equivalent
-     * to {@code buildFrom} except that the {@code FileDescriptorProto} is
-     * encoded in protocol buffer wire format.
-     */
-    public static void internalBuildGeneratedFileFrom(
-        final String[] descriptorDataParts,
-        final FileDescriptor[] dependencies,
-        final InternalDescriptorAssigner descriptorAssigner) {
-      // Hack:  We can't embed a raw byte array inside generated Java code
-      //   (at least, not efficiently), but we can embed Strings.  So, the
-      //   protocol compiler embeds the FileDescriptorProto as a giant
-      //   string literal which is passed to this function to construct the
-      //   file's FileDescriptor.  The string literal contains only 8-bit
-      //   characters, each one representing a byte of the FileDescriptorProto's
-      //   serialized form.  So, if we convert it to bytes in ISO-8859-1, we
-      //   should get the original bytes that we want.
-
-      // descriptorData may contain multiple strings in order to get around the
-      // Java 64k string literal limit.
-      StringBuilder descriptorData = new StringBuilder();
-      for (String part : descriptorDataParts) {
-        descriptorData.append(part);
-      }
-
-      final byte[] descriptorBytes;
-      descriptorBytes = descriptorData.toString().getBytes(Internal.ISO_8859_1);
-
-      FileDescriptorProto proto;
-      try {
-        proto = FileDescriptorProto.parseFrom(descriptorBytes);
-      } catch (InvalidProtocolBufferException e) {
-        throw new IllegalArgumentException(
-          "Failed to parse protocol buffer descriptor for generated code.", e);
-      }
-
-      final FileDescriptor result;
-      try {
-        // When building descriptors for generated code, we allow unknown
-        // dependencies by default.
-        result = buildFrom(proto, dependencies, true);
-      } catch (DescriptorValidationException e) {
-        throw new IllegalArgumentException(
-          "Invalid embedded descriptor for \"" + proto.getName() + "\".", e);
-      }
-
-      final ExtensionRegistry registry =
-          descriptorAssigner.assignDescriptors(result);
-
-      if (registry != null) {
-        // We must re-parse the proto using the registry.
-        try {
-          proto = FileDescriptorProto.parseFrom(descriptorBytes, registry);
-        } catch (InvalidProtocolBufferException e) {
-          throw new IllegalArgumentException(
-            "Failed to parse protocol buffer descriptor for generated code.",
-            e);
-        }
-
-        result.setProto(proto);
-      }
-    }
-
-    /**
-     * This method is to be called by generated code only.  It uses Java
-     * reflection to load the dependencies' descriptors.
-     */
-    public static void internalBuildGeneratedFileFrom(
-        final String[] descriptorDataParts,
-        final Class<?> descriptorOuterClass,
-        final String[] dependencies,
-        final String[] dependencyFileNames,
-        final InternalDescriptorAssigner descriptorAssigner) {
-      List<FileDescriptor> descriptors = new ArrayList<FileDescriptor>();
-      for (int i = 0; i < dependencies.length; i++) {
-        try {
-          Class<?> clazz =
-              descriptorOuterClass.getClassLoader().loadClass(dependencies[i]);
-          descriptors.add(
-              (FileDescriptor) clazz.getField("descriptor").get(null));
-        } catch (Exception e) {
-          // We allow unknown dependencies by default. If a dependency cannot
-          // be found we only generate a warning.
-          logger.warning("Descriptors for \"" + dependencyFileNames[i] +
-              "\" can not be found.");
-        }
-      }
-      FileDescriptor[] descriptorArray = new FileDescriptor[descriptors.size()];
-      descriptors.toArray(descriptorArray);
-      internalBuildGeneratedFileFrom(
-          descriptorDataParts, descriptorArray, descriptorAssigner);
-    }
-
-    /**
-     * This method is to be called by generated code only.  It is used to
-     * update the FileDescriptorProto associated with the descriptor by
-     * parsing it again with the given ExtensionRegistry. This is needed to
-     * recognize custom options.
-     */
-    public static void internalUpdateFileDescriptor(
-        final FileDescriptor descriptor,
-        final ExtensionRegistry registry) {
-      ByteString bytes = descriptor.proto.toByteString();
-      FileDescriptorProto proto;
-      try {
-        proto = FileDescriptorProto.parseFrom(bytes, registry);
-      } catch (InvalidProtocolBufferException e) {
-        throw new IllegalArgumentException(
-          "Failed to parse protocol buffer descriptor for generated code.", e);
-      }
-      descriptor.setProto(proto);
-    }
-
-    /**
-     * This class should be used by generated code only.  When calling
-     * {@link FileDescriptor#internalBuildGeneratedFileFrom}, the caller
-     * provides a callback implementing this interface.  The callback is called
-     * after the FileDescriptor has been constructed, in order to assign all
-     * the global variables defined in the generated code which point at parts
-     * of the FileDescriptor.  The callback returns an ExtensionRegistry which
-     * contains any extensions which might be used in the descriptor -- that
-     * is, extensions of the various "Options" messages defined in
-     * descriptor.proto.  The callback may also return null to indicate that
-     * no extensions are used in the descriptor.
-     */
-    public interface InternalDescriptorAssigner {
-      ExtensionRegistry assignDescriptors(FileDescriptor root);
-    }
-
-    private FileDescriptorProto proto;
-    private final Descriptor[] messageTypes;
-    private final EnumDescriptor[] enumTypes;
-    private final ServiceDescriptor[] services;
-    private final FieldDescriptor[] extensions;
-    private final FileDescriptor[] dependencies;
-    private final FileDescriptor[] publicDependencies;
-    private final DescriptorPool pool;
-
-    private FileDescriptor(final FileDescriptorProto proto,
-                           final FileDescriptor[] dependencies,
-                           final DescriptorPool pool,
-                           boolean allowUnknownDependencies)
-                    throws DescriptorValidationException {
-      this.pool = pool;
-      this.proto = proto;
-      this.dependencies = dependencies.clone();
-      HashMap<String, FileDescriptor> nameToFileMap =
-          new HashMap<String, FileDescriptor>();
-      for (FileDescriptor file : dependencies) {
-        nameToFileMap.put(file.getName(), file);
-      }
-      List<FileDescriptor> publicDependencies = new ArrayList<FileDescriptor>();
-      for (int i = 0; i < proto.getPublicDependencyCount(); i++) {
-        int index = proto.getPublicDependency(i);
-        if (index < 0 || index >= proto.getDependencyCount()) {
-          throw new DescriptorValidationException(this,
-              "Invalid public dependency index.");
-        }
-        String name = proto.getDependency(index);
-        FileDescriptor file = nameToFileMap.get(name);
-        if (file == null) {
-          if (!allowUnknownDependencies) {
-            throw new DescriptorValidationException(this,
-                "Invalid public dependency: " + name);
-          }
-          // Ignore unknown dependencies.
-        } else {
-          publicDependencies.add(file);
-        }
-      }
-      this.publicDependencies = new FileDescriptor[publicDependencies.size()];
-      publicDependencies.toArray(this.publicDependencies);
-
-      pool.addPackage(getPackage(), this);
-
-      messageTypes = new Descriptor[proto.getMessageTypeCount()];
-      for (int i = 0; i < proto.getMessageTypeCount(); i++) {
-        messageTypes[i] =
-          new Descriptor(proto.getMessageType(i), this, null, i);
-      }
-
-      enumTypes = new EnumDescriptor[proto.getEnumTypeCount()];
-      for (int i = 0; i < proto.getEnumTypeCount(); i++) {
-        enumTypes[i] = new EnumDescriptor(proto.getEnumType(i), this, null, i);
-      }
-
-      services = new ServiceDescriptor[proto.getServiceCount()];
-      for (int i = 0; i < proto.getServiceCount(); i++) {
-        services[i] = new ServiceDescriptor(proto.getService(i), this, i);
-      }
-
-      extensions = new FieldDescriptor[proto.getExtensionCount()];
-      for (int i = 0; i < proto.getExtensionCount(); i++) {
-        extensions[i] = new FieldDescriptor(
-          proto.getExtension(i), this, null, i, true);
-      }
-    }
-
-    /**
-     * Create a placeholder FileDescriptor for a message Descriptor.
-     */
-    FileDescriptor(String packageName, Descriptor message)
-        throws DescriptorValidationException {
-      this.pool = new DescriptorPool(new FileDescriptor[0], true);
-      this.proto = FileDescriptorProto.newBuilder()
-          .setName(message.getFullName() + ".placeholder.proto")
-          .setPackage(packageName).addMessageType(message.toProto()).build();
-      this.dependencies = new FileDescriptor[0];
-      this.publicDependencies = new FileDescriptor[0];
-
-      messageTypes = new Descriptor[] {message};
-      enumTypes = new EnumDescriptor[0];
-      services = new ServiceDescriptor[0];
-      extensions = new FieldDescriptor[0];
-
-      pool.addPackage(packageName, this);
-      pool.addSymbol(message);
-    }
-
-    /** Look up and cross-link all field types, etc. */
-    private void crossLink() throws DescriptorValidationException {
-      for (final Descriptor messageType : messageTypes) {
-        messageType.crossLink();
-      }
-
-      for (final ServiceDescriptor service : services) {
-        service.crossLink();
-      }
-
-      for (final FieldDescriptor extension : extensions) {
-        extension.crossLink();
-      }
-    }
-
-    /**
-     * Replace our {@link FileDescriptorProto} with the given one, which is
-     * identical except that it might contain extensions that weren't present
-     * in the original.  This method is needed for bootstrapping when a file
-     * defines custom options.  The options may be defined in the file itself,
-     * so we can't actually parse them until we've constructed the descriptors,
-     * but to construct the descriptors we have to have parsed the descriptor
-     * protos.  So, we have to parse the descriptor protos a second time after
-     * constructing the descriptors.
-     */
-    private void setProto(final FileDescriptorProto proto) {
-      this.proto = proto;
-
-      for (int i = 0; i < messageTypes.length; i++) {
-        messageTypes[i].setProto(proto.getMessageType(i));
-      }
-
-      for (int i = 0; i < enumTypes.length; i++) {
-        enumTypes[i].setProto(proto.getEnumType(i));
-      }
-
-      for (int i = 0; i < services.length; i++) {
-        services[i].setProto(proto.getService(i));
-      }
-
-      for (int i = 0; i < extensions.length; i++) {
-        extensions[i].setProto(proto.getExtension(i));
-      }
-    }
-
-    boolean supportsUnknownEnumValue() {
-      return getSyntax() == Syntax.PROTO3;
-    }
-  }
-
-  // =================================================================
-
-  /** Describes a message type. */
-  public static final class Descriptor extends GenericDescriptor {
-    /**
-     * Get the index of this descriptor within its parent.  In other words,
-     * given a {@link FileDescriptor} {@code file}, the following is true:
-     * <pre>
-     *   for all i in [0, file.getMessageTypeCount()):
-     *     file.getMessageType(i).getIndex() == i
-     * </pre>
-     * Similarly, for a {@link Descriptor} {@code messageType}:
-     * <pre>
-     *   for all i in [0, messageType.getNestedTypeCount()):
-     *     messageType.getNestedType(i).getIndex() == i
-     * </pre>
-     */
-    public int getIndex() { return index; }
-
-    /** Convert the descriptor to its protocol message representation. */
-    @Override
-    public DescriptorProto toProto() {
-      return proto;
-    }
-
-    /** Get the type's unqualified name. */
-    @Override
-    public String getName() {
-      return proto.getName();
-    }
-
-    /**
-     * Get the type's fully-qualified name, within the proto language's
-     * namespace.  This differs from the Java name.  For example, given this
-     * {@code .proto}:
-     * <pre>
-     *   package foo.bar;
-     *   option java_package = "com.example.protos"
-     *   message Baz {}
-     * </pre>
-     * {@code Baz}'s full name is "foo.bar.Baz".
-     */
-    @Override
-    public String getFullName() {
-      return fullName;
-    }
-
-    /** Get the {@link FileDescriptor} containing this descriptor. */
-    @Override
-    public FileDescriptor getFile() {
-      return file;
-    }
-
-    /** If this is a nested type, get the outer descriptor, otherwise null. */
-    public Descriptor getContainingType() { return containingType; }
-
-    /** Get the {@code MessageOptions}, defined in {@code descriptor.proto}. */
-    public MessageOptions getOptions() { return proto.getOptions(); }
-
-    /** Get a list of this message type's fields. */
-    public List<FieldDescriptor> getFields() {
-      return Collections.unmodifiableList(Arrays.asList(fields));
-    }
-
-    /** Get a list of this message type's oneofs. */
-    public List<OneofDescriptor> getOneofs() {
-      return Collections.unmodifiableList(Arrays.asList(oneofs));
-    }
-
-    /** Get a list of this message type's extensions. */
-    public List<FieldDescriptor> getExtensions() {
-      return Collections.unmodifiableList(Arrays.asList(extensions));
-    }
-
-    /** Get a list of message types nested within this one. */
-    public List<Descriptor> getNestedTypes() {
-      return Collections.unmodifiableList(Arrays.asList(nestedTypes));
-    }
-
-    /** Get a list of enum types nested within this one. */
-    public List<EnumDescriptor> getEnumTypes() {
-      return Collections.unmodifiableList(Arrays.asList(enumTypes));
-    }
-
-    /** Determines if the given field number is an extension. */
-    public boolean isExtensionNumber(final int number) {
-      for (final DescriptorProto.ExtensionRange range :
-          proto.getExtensionRangeList()) {
-        if (range.getStart() <= number && number < range.getEnd()) {
-          return true;
-        }
-      }
-      return false;
-    }
-
-    /** Determines if the given field number is reserved. */
-    public boolean isReservedNumber(final int number) {
-      for (final DescriptorProto.ReservedRange range :
-          proto.getReservedRangeList()) {
-        if (range.getStart() <= number && number < range.getEnd()) {
-          return true;
-        }
-      }
-      return false;
-    }
-
-    /** Determines if the given field name is reserved. */
-    public boolean isReservedName(final String name) {
-      if (name == null) {
-        throw new NullPointerException();
-      }
-      for (final String reservedName : proto.getReservedNameList()) {
-        if (reservedName.equals(name)) {
-          return true;
-        }
-      }
-      return false;
-    }
-
-    /**
-     * Indicates whether the message can be extended.  That is, whether it has
-     * any "extensions x to y" ranges declared on it.
-     */
-    public boolean isExtendable() {
-      return proto.getExtensionRangeList().size() != 0;
-    }
-
-    /**
-     * Finds a field by name.
-     * @param name The unqualified name of the field (e.g. "foo").
-     * @return The field's descriptor, or {@code null} if not found.
-     */
-    public FieldDescriptor findFieldByName(final String name) {
-      final GenericDescriptor result =
-          file.pool.findSymbol(fullName + '.' + name);
-      if (result != null && result instanceof FieldDescriptor) {
-        return (FieldDescriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    /**
-     * Finds a field by field number.
-     * @param number The field number within this message type.
-     * @return The field's descriptor, or {@code null} if not found.
-     */
-    public FieldDescriptor findFieldByNumber(final int number) {
-      return file.pool.fieldsByNumber.get(
-        new DescriptorPool.DescriptorIntPair(this, number));
-    }
-
-    /**
-     * Finds a nested message type by name.
-     * @param name The unqualified name of the nested type (e.g. "Foo").
-     * @return The types's descriptor, or {@code null} if not found.
-     */
-    public Descriptor findNestedTypeByName(final String name) {
-      final GenericDescriptor result =
-          file.pool.findSymbol(fullName + '.' + name);
-      if (result != null && result instanceof Descriptor) {
-        return (Descriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    /**
-     * Finds a nested enum type by name.
-     * @param name The unqualified name of the nested type (e.g. "Foo").
-     * @return The types's descriptor, or {@code null} if not found.
-     */
-    public EnumDescriptor findEnumTypeByName(final String name) {
-      final GenericDescriptor result =
-          file.pool.findSymbol(fullName + '.' + name);
-      if (result != null && result instanceof EnumDescriptor) {
-        return (EnumDescriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    private final int index;
-    private DescriptorProto proto;
-    private final String fullName;
-    private final FileDescriptor file;
-    private final Descriptor containingType;
-    private final Descriptor[] nestedTypes;
-    private final EnumDescriptor[] enumTypes;
-    private final FieldDescriptor[] fields;
-    private final FieldDescriptor[] extensions;
-    private final OneofDescriptor[] oneofs;
-
-    // Used to create a placeholder when the type cannot be found.
-    Descriptor(final String fullname) throws DescriptorValidationException {
-      String name = fullname;
-      String packageName = "";
-      int pos = fullname.lastIndexOf('.');
-      if (pos != -1) {
-        name = fullname.substring(pos + 1);
-        packageName = fullname.substring(0, pos);
-      }
-      this.index = 0;
-      this.proto = DescriptorProto.newBuilder().setName(name).addExtensionRange(
-          DescriptorProto.ExtensionRange.newBuilder().setStart(1)
-          .setEnd(536870912).build()).build();
-      this.fullName = fullname;
-      this.containingType = null;
-
-      this.nestedTypes = new Descriptor[0];
-      this.enumTypes = new EnumDescriptor[0];
-      this.fields = new FieldDescriptor[0];
-      this.extensions = new FieldDescriptor[0];
-      this.oneofs = new OneofDescriptor[0];
-
-      // Create a placeholder FileDescriptor to hold this message.
-      this.file = new FileDescriptor(packageName, this);
-    }
-
-    private Descriptor(final DescriptorProto proto,
-                       final FileDescriptor file,
-                       final Descriptor parent,
-                       final int index)
-                throws DescriptorValidationException {
-      this.index = index;
-      this.proto = proto;
-      fullName = computeFullName(file, parent, proto.getName());
-      this.file = file;
-      containingType = parent;
-
-      oneofs = new OneofDescriptor[proto.getOneofDeclCount()];
-      for (int i = 0; i < proto.getOneofDeclCount(); i++) {
-        oneofs[i] = new OneofDescriptor(
-          proto.getOneofDecl(i), file, this, i);
-      }
-
-      nestedTypes = new Descriptor[proto.getNestedTypeCount()];
-      for (int i = 0; i < proto.getNestedTypeCount(); i++) {
-        nestedTypes[i] = new Descriptor(
-          proto.getNestedType(i), file, this, i);
-      }
-
-      enumTypes = new EnumDescriptor[proto.getEnumTypeCount()];
-      for (int i = 0; i < proto.getEnumTypeCount(); i++) {
-        enumTypes[i] = new EnumDescriptor(
-          proto.getEnumType(i), file, this, i);
-      }
-
-      fields = new FieldDescriptor[proto.getFieldCount()];
-      for (int i = 0; i < proto.getFieldCount(); i++) {
-        fields[i] = new FieldDescriptor(
-          proto.getField(i), file, this, i, false);
-      }
-
-      extensions = new FieldDescriptor[proto.getExtensionCount()];
-      for (int i = 0; i < proto.getExtensionCount(); i++) {
-        extensions[i] = new FieldDescriptor(
-          proto.getExtension(i), file, this, i, true);
-      }
-
-      for (int i = 0; i < proto.getOneofDeclCount(); i++) {
-        oneofs[i].fields = new FieldDescriptor[oneofs[i].getFieldCount()];
-        oneofs[i].fieldCount = 0;
-      }
-      for (int i = 0; i < proto.getFieldCount(); i++) {
-        OneofDescriptor oneofDescriptor = fields[i].getContainingOneof();
-        if (oneofDescriptor != null) {
-          oneofDescriptor.fields[oneofDescriptor.fieldCount++] = fields[i];
-        }
-      }
-
-      file.pool.addSymbol(this);
-    }
-
-    /** Look up and cross-link all field types, etc. */
-    private void crossLink() throws DescriptorValidationException {
-      for (final Descriptor nestedType : nestedTypes) {
-        nestedType.crossLink();
-      }
-
-      for (final FieldDescriptor field : fields) {
-        field.crossLink();
-      }
-
-      for (final FieldDescriptor extension : extensions) {
-        extension.crossLink();
-      }
-    }
-
-    /** See {@link FileDescriptor#setProto}. */
-    private void setProto(final DescriptorProto proto) {
-      this.proto = proto;
-
-      for (int i = 0; i < nestedTypes.length; i++) {
-        nestedTypes[i].setProto(proto.getNestedType(i));
-      }
-
-      for (int i = 0; i < oneofs.length; i++) {
-        oneofs[i].setProto(proto.getOneofDecl(i));
-      }
-
-      for (int i = 0; i < enumTypes.length; i++) {
-        enumTypes[i].setProto(proto.getEnumType(i));
-      }
-
-      for (int i = 0; i < fields.length; i++) {
-        fields[i].setProto(proto.getField(i));
-      }
-
-      for (int i = 0; i < extensions.length; i++) {
-        extensions[i].setProto(proto.getExtension(i));
-      }
-    }
-  }
-
-  // =================================================================
-
-  /** Describes a field of a message type. */
-  public static final class FieldDescriptor
-        extends GenericDescriptor
-        implements Comparable<FieldDescriptor>,
-                 FieldSet.FieldDescriptorLite<FieldDescriptor> {
-    /**
-     * Get the index of this descriptor within its parent.
-     * @see Descriptors.Descriptor#getIndex()
-     */
-    public int getIndex() { return index; }
-
-    /** Convert the descriptor to its protocol message representation. */
-    @Override
-    public FieldDescriptorProto toProto() {
-      return proto;
-    }
-
-    /** Get the field's unqualified name. */
-    @Override
-    public String getName() {
-      return proto.getName();
-    }
-
-    /** Get the field's number. */
-    @Override
-    public int getNumber() {
-      return proto.getNumber();
-    }
-
-    /**
-     * Get the field's fully-qualified name.
-     * @see Descriptors.Descriptor#getFullName()
-     */
-    @Override
-    public String getFullName() {
-      return fullName;
-    }
-
-    /** Get the JSON name of this field. */
-    public String getJsonName() {
-      return jsonName;
-    }
-
-    /**
-     * Get the field's java type.  This is just for convenience.  Every
-     * {@code FieldDescriptorProto.Type} maps to exactly one Java type.
-     */
-    public JavaType getJavaType() { return type.getJavaType(); }
-
-    /** For internal use only. */
-    @Override
-    public WireFormat.JavaType getLiteJavaType() {
-      return getLiteType().getJavaType();
-    }
-
-    /** Get the {@code FileDescriptor} containing this descriptor. */
-    @Override
-    public FileDescriptor getFile() {
-      return file;
-    }
-
-    /** Get the field's declared type. */
-    public Type getType() { return type; }
-
-    /** For internal use only. */
-    @Override
-    public WireFormat.FieldType getLiteType() {
-      return table[type.ordinal()];
-    }
-
-    /** For internal use only. */
-    public boolean needsUtf8Check() {
-      if (type != Type.STRING) {
-        return false;
-      }
-      if (getContainingType().getOptions().getMapEntry()) {
-        // Always enforce strict UTF-8 checking for map fields.
-        return true;
-      }
-      if (getFile().getSyntax() == Syntax.PROTO3) {
-        return true;
-      }
-      return getFile().getOptions().getJavaStringCheckUtf8();
-    }
-
-    public boolean isMapField() {
-      return getType() == Type.MESSAGE && isRepeated()
-          && getMessageType().getOptions().getMapEntry();
-    }
-
-    // I'm pretty sure values() constructs a new array every time, since there
-    // is nothing stopping the caller from mutating the array.  Therefore we
-    // make a static copy here.
-    private static final WireFormat.FieldType[] table =
-        WireFormat.FieldType.values();
-
-    /** Is this field declared required? */
-    public boolean isRequired() {
-      return proto.getLabel() == FieldDescriptorProto.Label.LABEL_REQUIRED;
-    }
-
-    /** Is this field declared optional? */
-    public boolean isOptional() {
-      return proto.getLabel() == FieldDescriptorProto.Label.LABEL_OPTIONAL;
-    }
-
-    /** Is this field declared repeated? */
-    @Override
-    public boolean isRepeated() {
-      return proto.getLabel() == FieldDescriptorProto.Label.LABEL_REPEATED;
-    }
-
-    /** Does this field have the {@code [packed = true]} option or is this field
-     *  packable in proto3 and not explicitly setted to unpacked?
-     */
-    @Override
-    public boolean isPacked() {
-      if (!isPackable()) {
-        return false;
-      }
-      if (getFile().getSyntax() == FileDescriptor.Syntax.PROTO2) {
-        return getOptions().getPacked();
-      } else {
-        return !getOptions().hasPacked() || getOptions().getPacked();
-      }
-    }
-
-    /** Can this field be packed? i.e. is it a repeated primitive field? */
-    public boolean isPackable() {
-      return isRepeated() && getLiteType().isPackable();
-    }
-
-    /** Returns true if the field had an explicitly-defined default value. */
-    public boolean hasDefaultValue() { return proto.hasDefaultValue(); }
-
-    /**
-     * Returns the field's default value.  Valid for all types except for
-     * messages and groups.  For all other types, the object returned is of
-     * the same class that would returned by Message.getField(this).
-     */
-    public Object getDefaultValue() {
-      if (getJavaType() == JavaType.MESSAGE) {
-        throw new UnsupportedOperationException(
-          "FieldDescriptor.getDefaultValue() called on an embedded message " +
-          "field.");
-      }
-      return defaultValue;
-    }
-
-    /** Get the {@code FieldOptions}, defined in {@code descriptor.proto}. */
-    public FieldOptions getOptions() { return proto.getOptions(); }
-
-    /** Is this field an extension? */
-    public boolean isExtension() { return proto.hasExtendee(); }
-
-    /**
-     * Get the field's containing type. For extensions, this is the type being
-     * extended, not the location where the extension was defined.  See
-     * {@link #getExtensionScope()}.
-     */
-    public Descriptor getContainingType() { return containingType; }
-
-    /** Get the field's containing oneof. */
-    public OneofDescriptor getContainingOneof() { return containingOneof; }
-
-    /**
-     * For extensions defined nested within message types, gets the outer
-     * type.  Not valid for non-extension fields.  For example, consider
-     * this {@code .proto} file:
-     * <pre>
-     *   message Foo {
-     *     extensions 1000 to max;
-     *   }
-     *   extend Foo {
-     *     optional int32 baz = 1234;
-     *   }
-     *   message Bar {
-     *     extend Foo {
-     *       optional int32 qux = 4321;
-     *     }
-     *   }
-     * </pre>
-     * Both {@code baz}'s and {@code qux}'s containing type is {@code Foo}.
-     * However, {@code baz}'s extension scope is {@code null} while
-     * {@code qux}'s extension scope is {@code Bar}.
-     */
-    public Descriptor getExtensionScope() {
-      if (!isExtension()) {
-        throw new UnsupportedOperationException(
-          "This field is not an extension.");
-      }
-      return extensionScope;
-    }
-
-    /** For embedded message and group fields, gets the field's type. */
-    public Descriptor getMessageType() {
-      if (getJavaType() != JavaType.MESSAGE) {
-        throw new UnsupportedOperationException(
-          "This field is not of message type.");
-      }
-      return messageType;
-    }
-
-    /** For enum fields, gets the field's type. */
-    @Override
-    public EnumDescriptor getEnumType() {
-      if (getJavaType() != JavaType.ENUM) {
-        throw new UnsupportedOperationException(
-          "This field is not of enum type.");
-      }
-      return enumType;
-    }
-
-    /**
-     * Compare with another {@code FieldDescriptor}.  This orders fields in
-     * "canonical" order, which simply means ascending order by field number.
-     * {@code other} must be a field of the same type -- i.e.
-     * {@code getContainingType()} must return the same {@code Descriptor} for
-     * both fields.
-     *
-     * @return negative, zero, or positive if {@code this} is less than,
-     *         equal to, or greater than {@code other}, respectively.
-     */
-    @Override
-    public int compareTo(final FieldDescriptor other) {
-      if (other.containingType != containingType) {
-        throw new IllegalArgumentException(
-          "FieldDescriptors can only be compared to other FieldDescriptors " +
-          "for fields of the same message type.");
-      }
-      return getNumber() - other.getNumber();
-    }
-
-    @Override
-    public String toString() {
-      return getFullName();
-    }
-
-    private final int index;
-
-    private FieldDescriptorProto proto;
-    private final String fullName;
-    private final String jsonName;
-    private final FileDescriptor file;
-    private final Descriptor extensionScope;
-
-    // Possibly initialized during cross-linking.
-    private Type type;
-    private Descriptor containingType;
-    private Descriptor messageType;
-    private OneofDescriptor containingOneof;
-    private EnumDescriptor enumType;
-    private Object defaultValue;
-
-    public enum Type {
-      DOUBLE  (JavaType.DOUBLE     ),
-      FLOAT   (JavaType.FLOAT      ),
-      INT64   (JavaType.LONG       ),
-      UINT64  (JavaType.LONG       ),
-      INT32   (JavaType.INT        ),
-      FIXED64 (JavaType.LONG       ),
-      FIXED32 (JavaType.INT        ),
-      BOOL    (JavaType.BOOLEAN    ),
-      STRING  (JavaType.STRING     ),
-      GROUP   (JavaType.MESSAGE    ),
-      MESSAGE (JavaType.MESSAGE    ),
-      BYTES   (JavaType.BYTE_STRING),
-      UINT32  (JavaType.INT        ),
-      ENUM    (JavaType.ENUM       ),
-      SFIXED32(JavaType.INT        ),
-      SFIXED64(JavaType.LONG       ),
-      SINT32  (JavaType.INT        ),
-      SINT64  (JavaType.LONG       );
-
-      Type(final JavaType javaType) {
-        this.javaType = javaType;
-      }
-
-      private JavaType javaType;
-
-      public FieldDescriptorProto.Type toProto() {
-        return FieldDescriptorProto.Type.forNumber(ordinal() + 1);
-      }
-      public JavaType getJavaType() { return javaType; }
-
-      public static Type valueOf(final FieldDescriptorProto.Type type) {
-        return values()[type.getNumber() - 1];
-      }
-    }
-
-    static {
-      // Refuse to init if someone added a new declared type.
-      if (Type.values().length != FieldDescriptorProto.Type.values().length) {
-        throw new RuntimeException(""
-            + "descriptor.proto has a new declared type but Descriptors.java "
-            + "wasn't updated.");
-      }
-    }
-
-    public enum JavaType {
-      INT(0),
-      LONG(0L),
-      FLOAT(0F),
-      DOUBLE(0D),
-      BOOLEAN(false),
-      STRING(""),
-      BYTE_STRING(ByteString.EMPTY),
-      ENUM(null),
-      MESSAGE(null);
-
-      JavaType(final Object defaultDefault) {
-        this.defaultDefault = defaultDefault;
-      }
-
-      /**
-       * The default default value for fields of this type, if it's a primitive
-       * type.  This is meant for use inside this file only, hence is private.
-       */
-      private final Object defaultDefault;
-    }
-
-    // This method should match exactly with the ToJsonName() function in C++
-    // descriptor.cc.
-    private static String fieldNameToJsonName(String name) {
-      StringBuilder result = new StringBuilder(name.length());
-      boolean isNextUpperCase = false;
-      for (int i = 0; i < name.length(); i++) {
-        Character ch = name.charAt(i);
-        if (ch == '_') {
-          isNextUpperCase = true;
-        } else if (isNextUpperCase) {
-          result.append(Character.toUpperCase(ch));
-          isNextUpperCase = false;
-        } else {
-          result.append(ch);
-        }
-      }
-      return result.toString();
-    }
-
-    private FieldDescriptor(final FieldDescriptorProto proto,
-                            final FileDescriptor file,
-                            final Descriptor parent,
-                            final int index,
-                            final boolean isExtension)
-                     throws DescriptorValidationException {
-      this.index = index;
-      this.proto = proto;
-      fullName = computeFullName(file, parent, proto.getName());
-      this.file = file;
-      if (proto.hasJsonName()) {
-        jsonName = proto.getJsonName();
-      } else {
-        jsonName = fieldNameToJsonName(proto.getName());
-      }
-
-      if (proto.hasType()) {
-        type = Type.valueOf(proto.getType());
-      }
-
-      if (getNumber() <= 0) {
-        throw new DescriptorValidationException(this,
-          "Field numbers must be positive integers.");
-      }
-
-      if (isExtension) {
-        if (!proto.hasExtendee()) {
-          throw new DescriptorValidationException(this,
-            "FieldDescriptorProto.extendee not set for extension field.");
-        }
-        containingType = null;  // Will be filled in when cross-linking
-        if (parent != null) {
-          extensionScope = parent;
-        } else {
-          extensionScope = null;
-        }
-
-        if (proto.hasOneofIndex()) {
-          throw new DescriptorValidationException(this,
-            "FieldDescriptorProto.oneof_index set for extension field.");
-        }
-        containingOneof = null;
-      } else {
-        if (proto.hasExtendee()) {
-          throw new DescriptorValidationException(this,
-            "FieldDescriptorProto.extendee set for non-extension field.");
-        }
-        containingType = parent;
-
-        if (proto.hasOneofIndex()) {
-          if (proto.getOneofIndex() < 0 ||
-              proto.getOneofIndex() >= parent.toProto().getOneofDeclCount()) {
-            throw new DescriptorValidationException(this,
-              "FieldDescriptorProto.oneof_index is out of range for type "
-              + parent.getName());
-          }
-          containingOneof = parent.getOneofs().get(proto.getOneofIndex());
-          containingOneof.fieldCount++;
-        } else {
-          containingOneof = null;
-        }
-        extensionScope = null;
-      }
-
-      file.pool.addSymbol(this);
-    }
-
-    /** Look up and cross-link all field types, etc. */
-    private void crossLink() throws DescriptorValidationException {
-      if (proto.hasExtendee()) {
-        final GenericDescriptor extendee =
-          file.pool.lookupSymbol(proto.getExtendee(), this,
-              DescriptorPool.SearchFilter.TYPES_ONLY);
-        if (!(extendee instanceof Descriptor)) {
-          throw new DescriptorValidationException(this,
-              '\"' + proto.getExtendee() + "\" is not a message type.");
-        }
-        containingType = (Descriptor)extendee;
-
-        if (!getContainingType().isExtensionNumber(getNumber())) {
-          throw new DescriptorValidationException(this,
-              '\"' + getContainingType().getFullName() +
-              "\" does not declare " + getNumber() +
-              " as an extension number.");
-        }
-      }
-
-      if (proto.hasTypeName()) {
-        final GenericDescriptor typeDescriptor =
-          file.pool.lookupSymbol(proto.getTypeName(), this,
-              DescriptorPool.SearchFilter.TYPES_ONLY);
-
-        if (!proto.hasType()) {
-          // Choose field type based on symbol.
-          if (typeDescriptor instanceof Descriptor) {
-            type = Type.MESSAGE;
-          } else if (typeDescriptor instanceof EnumDescriptor) {
-            type = Type.ENUM;
-          } else {
-            throw new DescriptorValidationException(this,
-                '\"' + proto.getTypeName() + "\" is not a type.");
-          }
-        }
-
-        if (getJavaType() == JavaType.MESSAGE) {
-          if (!(typeDescriptor instanceof Descriptor)) {
-            throw new DescriptorValidationException(this,
-                '\"' + proto.getTypeName() + "\" is not a message type.");
-          }
-          messageType = (Descriptor)typeDescriptor;
-
-          if (proto.hasDefaultValue()) {
-            throw new DescriptorValidationException(this,
-              "Messages can't have default values.");
-          }
-        } else if (getJavaType() == JavaType.ENUM) {
-          if (!(typeDescriptor instanceof EnumDescriptor)) {
-            throw new DescriptorValidationException(this,
-                '\"' + proto.getTypeName() + "\" is not an enum type.");
-          }
-          enumType = (EnumDescriptor)typeDescriptor;
-        } else {
-          throw new DescriptorValidationException(this,
-            "Field with primitive type has type_name.");
-        }
-      } else {
-        if (getJavaType() == JavaType.MESSAGE ||
-            getJavaType() == JavaType.ENUM) {
-          throw new DescriptorValidationException(this,
-            "Field with message or enum type missing type_name.");
-        }
-      }
-
-      // Only repeated primitive fields may be packed.
-      if (proto.getOptions().getPacked() && !isPackable()) {
-        throw new DescriptorValidationException(this,
-          "[packed = true] can only be specified for repeated primitive " +
-          "fields.");
-      }
-
-      // We don't attempt to parse the default value until here because for
-      // enums we need the enum type's descriptor.
-      if (proto.hasDefaultValue()) {
-        if (isRepeated()) {
-          throw new DescriptorValidationException(this,
-            "Repeated fields cannot have default values.");
-        }
-
-        try {
-          switch (getType()) {
-            case INT32:
-            case SINT32:
-            case SFIXED32:
-              defaultValue = TextFormat.parseInt32(proto.getDefaultValue());
-              break;
-            case UINT32:
-            case FIXED32:
-              defaultValue = TextFormat.parseUInt32(proto.getDefaultValue());
-              break;
-            case INT64:
-            case SINT64:
-            case SFIXED64:
-              defaultValue = TextFormat.parseInt64(proto.getDefaultValue());
-              break;
-            case UINT64:
-            case FIXED64:
-              defaultValue = TextFormat.parseUInt64(proto.getDefaultValue());
-              break;
-            case FLOAT:
-              if (proto.getDefaultValue().equals("inf")) {
-                defaultValue = Float.POSITIVE_INFINITY;
-              } else if (proto.getDefaultValue().equals("-inf")) {
-                defaultValue = Float.NEGATIVE_INFINITY;
-              } else if (proto.getDefaultValue().equals("nan")) {
-                defaultValue = Float.NaN;
-              } else {
-                defaultValue = Float.valueOf(proto.getDefaultValue());
-              }
-              break;
-            case DOUBLE:
-              if (proto.getDefaultValue().equals("inf")) {
-                defaultValue = Double.POSITIVE_INFINITY;
-              } else if (proto.getDefaultValue().equals("-inf")) {
-                defaultValue = Double.NEGATIVE_INFINITY;
-              } else if (proto.getDefaultValue().equals("nan")) {
-                defaultValue = Double.NaN;
-              } else {
-                defaultValue = Double.valueOf(proto.getDefaultValue());
-              }
-              break;
-            case BOOL:
-              defaultValue = Boolean.valueOf(proto.getDefaultValue());
-              break;
-            case STRING:
-              defaultValue = proto.getDefaultValue();
-              break;
-            case BYTES:
-              try {
-                defaultValue =
-                  TextFormat.unescapeBytes(proto.getDefaultValue());
-              } catch (TextFormat.InvalidEscapeSequenceException e) {
-                throw new DescriptorValidationException(this,
-                  "Couldn't parse default value: " + e.getMessage(), e);
-              }
-              break;
-            case ENUM:
-              defaultValue = enumType.findValueByName(proto.getDefaultValue());
-              if (defaultValue == null) {
-                throw new DescriptorValidationException(this,
-                  "Unknown enum default value: \"" +
-                  proto.getDefaultValue() + '\"');
-              }
-              break;
-            case MESSAGE:
-            case GROUP:
-              throw new DescriptorValidationException(this,
-                "Message type had default value.");
-          }
-        } catch (NumberFormatException e) {
-          throw new DescriptorValidationException(this,
-              "Could not parse default value: \"" +
-              proto.getDefaultValue() + '\"', e);
-        }
-      } else {
-        // Determine the default default for this field.
-        if (isRepeated()) {
-          defaultValue = Collections.emptyList();
-        } else {
-          switch (getJavaType()) {
-            case ENUM:
-              // We guarantee elsewhere that an enum type always has at least
-              // one possible value.
-              defaultValue = enumType.getValues().get(0);
-              break;
-            case MESSAGE:
-              defaultValue = null;
-              break;
-            default:
-              defaultValue = getJavaType().defaultDefault;
-              break;
-          }
-        }
-      }
-
-      if (!isExtension()) {
-        file.pool.addFieldByNumber(this);
-      }
-
-      if (containingType != null &&
-          containingType.getOptions().getMessageSetWireFormat()) {
-        if (isExtension()) {
-          if (!isOptional() || getType() != Type.MESSAGE) {
-            throw new DescriptorValidationException(this,
-              "Extensions of MessageSets must be optional messages.");
-          }
-        } else {
-          throw new DescriptorValidationException(this,
-            "MessageSets cannot have fields, only extensions.");
-        }
-      }
-    }
-
-    /** See {@link FileDescriptor#setProto}. */
-    private void setProto(final FieldDescriptorProto proto) {
-      this.proto = proto;
-    }
-
-    /**
-     * For internal use only.  This is to satisfy the FieldDescriptorLite
-     * interface.
-     */
-    @Override
-    public MessageLite.Builder internalMergeFrom(MessageLite.Builder to, MessageLite from) {
-      // FieldDescriptors are only used with non-lite messages so we can just
-      // down-cast and call mergeFrom directly.
-      return ((Message.Builder) to).mergeFrom((Message) from);
-    }
-
-  }
-
-  // =================================================================
-
-  /** Describes an enum type. */
-  public static final class EnumDescriptor extends GenericDescriptor
-      implements Internal.EnumLiteMap<EnumValueDescriptor> {
-    /**
-     * Get the index of this descriptor within its parent.
-     * @see Descriptors.Descriptor#getIndex()
-     */
-    public int getIndex() { return index; }
-
-    /** Convert the descriptor to its protocol message representation. */
-    @Override
-    public EnumDescriptorProto toProto() {
-      return proto;
-    }
-
-    /** Get the type's unqualified name. */
-    @Override
-    public String getName() {
-      return proto.getName();
-    }
-
-    /**
-     * Get the type's fully-qualified name.
-     * @see Descriptors.Descriptor#getFullName()
-     */
-    @Override
-    public String getFullName() {
-      return fullName;
-    }
-
-    /** Get the {@link FileDescriptor} containing this descriptor. */
-    @Override
-    public FileDescriptor getFile() {
-      return file;
-    }
-
-    /** If this is a nested type, get the outer descriptor, otherwise null. */
-    public Descriptor getContainingType() { return containingType; }
-
-    /** Get the {@code EnumOptions}, defined in {@code descriptor.proto}. */
-    public EnumOptions getOptions() { return proto.getOptions(); }
-
-    /** Get a list of defined values for this enum. */
-    public List<EnumValueDescriptor> getValues() {
-      return Collections.unmodifiableList(Arrays.asList(values));
-    }
-
-    /**
-     * Find an enum value by name.
-     * @param name The unqualified name of the value (e.g. "FOO").
-     * @return the value's descriptor, or {@code null} if not found.
-     */
-    public EnumValueDescriptor findValueByName(final String name) {
-      final GenericDescriptor result =
-          file.pool.findSymbol(fullName + '.' + name);
-      if (result != null && result instanceof EnumValueDescriptor) {
-        return (EnumValueDescriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    /**
-     * Find an enum value by number.  If multiple enum values have the same
-     * number, this returns the first defined value with that number.
-     * @param number The value's number.
-     * @return the value's descriptor, or {@code null} if not found.
-     */
-    @Override
-    public EnumValueDescriptor findValueByNumber(final int number) {
-      return file.pool.enumValuesByNumber.get(
-        new DescriptorPool.DescriptorIntPair(this, number));
-    }
-
-    /**
-     * Get the enum value for a number. If no enum value has this number,
-     * construct an EnumValueDescriptor for it.
-     */
-    public EnumValueDescriptor findValueByNumberCreatingIfUnknown(final int number) {
-      EnumValueDescriptor result = findValueByNumber(number);
-      if (result != null) {
-        return result;
-      }
-      // The number represents an unknown enum value.
-      synchronized (this) {
-        // Descriptors are compared by object identity so for the same number
-        // we need to return the same EnumValueDescriptor object. This means
-        // we have to store created EnumValueDescriptors. However, as there
-        // are potentially 2G unknown enum values, storing all of these
-        // objects persistently will consume lots of memory for long-running
-        // services and it's also unnecessary as not many EnumValueDescriptors
-        // will be used at the same time.
-        //
-        // To solve the problem we take advantage of Java's weak references and
-        // rely on gc to release unused descriptors.
-        //
-        // Here is how it works:
-        //   * We store unknown EnumValueDescriptors in a WeakHashMap with the
-        //     value being a weak reference to the descriptor.
-        //   * The descriptor holds a strong reference to the key so as long
-        //     as the EnumValueDescriptor is in use, the key will be there
-        //     and the corresponding map entry will be there. Following-up
-        //     queries with the same number will return the same descriptor.
-        //   * If the user no longer uses an unknown EnumValueDescriptor,
-        //     it will be gc-ed since we only hold a weak reference to it in
-        //     the map. The key in the corresponding map entry will also be
-        //     gc-ed as the only strong reference to it is in the descriptor
-        //     which is just gc-ed. With the key being gone WeakHashMap will
-        //     then remove the whole entry. This way unknown descriptors will
-        //     be freed automatically and we don't need to do anything to
-        //     clean-up unused map entries.
-
-        // Note: We must use "new Integer(number)" here because we don't want
-        // these Integer objects to be cached.
-        Integer key = new Integer(number);
-        WeakReference<EnumValueDescriptor> reference = unknownValues.get(key);
-        if (reference != null) {
-          result = reference.get();
-        }
-        if (result == null) {
-          result = new EnumValueDescriptor(file, this, key);
-          unknownValues.put(key, new WeakReference<EnumValueDescriptor>(result));
-        }
-      }
-      return result;
-    }
-
-    // Used in tests only.
-    int getUnknownEnumValueDescriptorCount() {
-      return unknownValues.size();
-    }
-
-    private final int index;
-    private EnumDescriptorProto proto;
-    private final String fullName;
-    private final FileDescriptor file;
-    private final Descriptor containingType;
-    private EnumValueDescriptor[] values;
-    private final WeakHashMap<Integer, WeakReference<EnumValueDescriptor>> unknownValues =
-        new WeakHashMap<Integer, WeakReference<EnumValueDescriptor>>();
-
-    private EnumDescriptor(final EnumDescriptorProto proto,
-                           final FileDescriptor file,
-                           final Descriptor parent,
-                           final int index)
-                    throws DescriptorValidationException {
-      this.index = index;
-      this.proto = proto;
-      fullName = computeFullName(file, parent, proto.getName());
-      this.file = file;
-      containingType = parent;
-
-      if (proto.getValueCount() == 0) {
-        // We cannot allow enums with no values because this would mean there
-        // would be no valid default value for fields of this type.
-        throw new DescriptorValidationException(this,
-          "Enums must contain at least one value.");
-      }
-
-      values = new EnumValueDescriptor[proto.getValueCount()];
-      for (int i = 0; i < proto.getValueCount(); i++) {
-        values[i] = new EnumValueDescriptor(
-          proto.getValue(i), file, this, i);
-      }
-
-      file.pool.addSymbol(this);
-    }
-
-    /** See {@link FileDescriptor#setProto}. */
-    private void setProto(final EnumDescriptorProto proto) {
-      this.proto = proto;
-
-      for (int i = 0; i < values.length; i++) {
-        values[i].setProto(proto.getValue(i));
-      }
-    }
-  }
-
-  // =================================================================
-
-  /**
-   * Describes one value within an enum type.  Note that multiple defined
-   * values may have the same number.  In generated Java code, all values
-   * with the same number after the first become aliases of the first.
-   * However, they still have independent EnumValueDescriptors.
-   */
-  public static final class EnumValueDescriptor extends GenericDescriptor
-      implements Internal.EnumLite {
-    /**
-     * Get the index of this descriptor within its parent.
-     * @see Descriptors.Descriptor#getIndex()
-     */
-    public int getIndex() { return index; }
-
-    /** Convert the descriptor to its protocol message representation. */
-    @Override
-    public EnumValueDescriptorProto toProto() {
-      return proto;
-    }
-
-    /** Get the value's unqualified name. */
-    @Override
-    public String getName() {
-      return proto.getName();
-    }
-
-    /** Get the value's number. */
-    @Override
-    public int getNumber() {
-      return proto.getNumber();
-    }
-
-    @Override
-    public String toString() { return proto.getName(); }
-
-    /**
-     * Get the value's fully-qualified name.
-     * @see Descriptors.Descriptor#getFullName()
-     */
-    @Override
-    public String getFullName() {
-      return fullName;
-    }
-
-    /** Get the {@link FileDescriptor} containing this descriptor. */
-    @Override
-    public FileDescriptor getFile() {
-      return file;
-    }
-
-    /** Get the value's enum type. */
-    public EnumDescriptor getType() { return type; }
-
-    /**
-     * Get the {@code EnumValueOptions}, defined in {@code descriptor.proto}.
-     */
-    public EnumValueOptions getOptions() { return proto.getOptions(); }
-
-    private final int index;
-    private EnumValueDescriptorProto proto;
-    private final String fullName;
-    private final FileDescriptor file;
-    private final EnumDescriptor type;
-
-    private EnumValueDescriptor(final EnumValueDescriptorProto proto,
-                                final FileDescriptor file,
-                                final EnumDescriptor parent,
-                                final int index)
-                         throws DescriptorValidationException {
-      this.index = index;
-      this.proto = proto;
-      this.file = file;
-      type = parent;
-
-      fullName = parent.getFullName() + '.' + proto.getName();
-
-      file.pool.addSymbol(this);
-      file.pool.addEnumValueByNumber(this);
-    }
-
-    private Integer number;
-    // Create an unknown enum value.
-    private EnumValueDescriptor(
-        final FileDescriptor file,
-        final EnumDescriptor parent,
-        final Integer number) {
-      String name = "UNKNOWN_ENUM_VALUE_" + parent.getName() + "_" + number;
-      EnumValueDescriptorProto proto = EnumValueDescriptorProto
-          .newBuilder().setName(name).setNumber(number).build();
-      this.index = -1;
-      this.proto = proto;
-      this.file = file;
-      this.type = parent;
-      this.fullName = parent.getFullName() + '.' + proto.getName();
-      this.number = number;
-
-      // Don't add this descriptor into pool.
-    }
-
-    /** See {@link FileDescriptor#setProto}. */
-    private void setProto(final EnumValueDescriptorProto proto) {
-      this.proto = proto;
-    }
-  }
-
-  // =================================================================
-
-  /** Describes a service type. */
-  public static final class ServiceDescriptor extends GenericDescriptor {
-    /**
-     * Get the index of this descriptor within its parent.
-     * * @see Descriptors.Descriptor#getIndex()
-     */
-    public int getIndex() { return index; }
-
-    /** Convert the descriptor to its protocol message representation. */
-    @Override
-    public ServiceDescriptorProto toProto() {
-      return proto;
-    }
-
-    /** Get the type's unqualified name. */
-    @Override
-    public String getName() {
-      return proto.getName();
-    }
-
-    /**
-     * Get the type's fully-qualified name.
-     * @see Descriptors.Descriptor#getFullName()
-     */
-    @Override
-    public String getFullName() {
-      return fullName;
-    }
-
-    /** Get the {@link FileDescriptor} containing this descriptor. */
-    @Override
-    public FileDescriptor getFile() {
-      return file;
-    }
-
-    /** Get the {@code ServiceOptions}, defined in {@code descriptor.proto}. */
-    public ServiceOptions getOptions() { return proto.getOptions(); }
-
-    /** Get a list of methods for this service. */
-    public List<MethodDescriptor> getMethods() {
-      return Collections.unmodifiableList(Arrays.asList(methods));
-    }
-
-    /**
-     * Find a method by name.
-     * @param name The unqualified name of the method (e.g. "Foo").
-     * @return the method's descriptor, or {@code null} if not found.
-     */
-    public MethodDescriptor findMethodByName(final String name) {
-      final GenericDescriptor result =
-          file.pool.findSymbol(fullName + '.' + name);
-      if (result != null && result instanceof MethodDescriptor) {
-        return (MethodDescriptor)result;
-      } else {
-        return null;
-      }
-    }
-
-    private final int index;
-    private ServiceDescriptorProto proto;
-    private final String fullName;
-    private final FileDescriptor file;
-    private MethodDescriptor[] methods;
-
-    private ServiceDescriptor(final ServiceDescriptorProto proto,
-                              final FileDescriptor file,
-                              final int index)
-                       throws DescriptorValidationException {
-      this.index = index;
-      this.proto = proto;
-      fullName = computeFullName(file, null, proto.getName());
-      this.file = file;
-
-      methods = new MethodDescriptor[proto.getMethodCount()];
-      for (int i = 0; i < proto.getMethodCount(); i++) {
-        methods[i] = new MethodDescriptor(
-          proto.getMethod(i), file, this, i);
-      }
-
-      file.pool.addSymbol(this);
-    }
-
-    private void crossLink() throws DescriptorValidationException {
-      for (final MethodDescriptor method : methods) {
-        method.crossLink();
-      }
-    }
-
-    /** See {@link FileDescriptor#setProto}. */
-    private void setProto(final ServiceDescriptorProto proto) {
-      this.proto = proto;
-
-      for (int i = 0; i < methods.length; i++) {
-        methods[i].setProto(proto.getMethod(i));
-      }
-    }
-  }
-
-  // =================================================================
-
-  /**
-   * Describes one method within a service type.
-   */
-  public static final class MethodDescriptor extends GenericDescriptor {
-    /**
-     * Get the index of this descriptor within its parent.
-     * * @see Descriptors.Descriptor#getIndex()
-     */
-    public int getIndex() { return index; }
-
-    /** Convert the descriptor to its protocol message representation. */
-    @Override
-    public MethodDescriptorProto toProto() {
-      return proto;
-    }
-
-    /** Get the method's unqualified name. */
-    @Override
-    public String getName() {
-      return proto.getName();
-    }
-
-    /**
-     * Get the method's fully-qualified name.
-     * @see Descriptors.Descriptor#getFullName()
-     */
-    @Override
-    public String getFullName() {
-      return fullName;
-    }
-
-    /** Get the {@link FileDescriptor} containing this descriptor. */
-    @Override
-    public FileDescriptor getFile() {
-      return file;
-    }
-
-    /** Get the method's service type. */
-    public ServiceDescriptor getService() { return service; }
-
-    /** Get the method's input type. */
-    public Descriptor getInputType() { return inputType; }
-
-    /** Get the method's output type. */
-    public Descriptor getOutputType() { return outputType; }
-
-    /**
-     * Get the {@code MethodOptions}, defined in {@code descriptor.proto}.
-     */
-    public MethodOptions getOptions() { return proto.getOptions(); }
-
-    private final int index;
-    private MethodDescriptorProto proto;
-    private final String fullName;
-    private final FileDescriptor file;
-    private final ServiceDescriptor service;
-
-    // Initialized during cross-linking.
-    private Descriptor inputType;
-    private Descriptor outputType;
-
-    private MethodDescriptor(final MethodDescriptorProto proto,
-                             final FileDescriptor file,
-                             final ServiceDescriptor parent,
-                             final int index)
-                      throws DescriptorValidationException {
-      this.index = index;
-      this.proto = proto;
-      this.file = file;
-      service = parent;
-
-      fullName = parent.getFullName() + '.' + proto.getName();
-
-      file.pool.addSymbol(this);
-    }
-
-    private void crossLink() throws DescriptorValidationException {
-      final GenericDescriptor input =
-        file.pool.lookupSymbol(proto.getInputType(), this,
-            DescriptorPool.SearchFilter.TYPES_ONLY);
-      if (!(input instanceof Descriptor)) {
-        throw new DescriptorValidationException(this,
-            '\"' + proto.getInputType() + "\" is not a message type.");
-      }
-      inputType = (Descriptor)input;
-
-      final GenericDescriptor output =
-        file.pool.lookupSymbol(proto.getOutputType(), this,
-            DescriptorPool.SearchFilter.TYPES_ONLY);
-      if (!(output instanceof Descriptor)) {
-        throw new DescriptorValidationException(this,
-            '\"' + proto.getOutputType() + "\" is not a message type.");
-      }
-      outputType = (Descriptor)output;
-    }
-
-    /** See {@link FileDescriptor#setProto}. */
-    private void setProto(final MethodDescriptorProto proto) {
-      this.proto = proto;
-    }
-  }
-
-  // =================================================================
-
-  private static String computeFullName(final FileDescriptor file,
-                                        final Descriptor parent,
-                                        final String name) {
-    if (parent != null) {
-      return parent.getFullName() + '.' + name;
-    } else if (file.getPackage().length() > 0) {
-      return file.getPackage() + '.' + name;
-    } else {
-      return name;
-    }
-  }
-
-  // =================================================================
-
-  /**
-   * All descriptors implement this to make it easier to implement tools like
-   * {@code DescriptorPool}.<p>
-   *
-   * This class is public so that the methods it exposes can be called from
-   * outside of this package. However, it should only be subclassed from
-   * nested classes of Descriptors.
-   */
-  public abstract static class GenericDescriptor {
-    public abstract Message toProto();
-    public abstract String getName();
-    public abstract String getFullName();
-    public abstract FileDescriptor getFile();
-  }
-
-  /**
-   * Thrown when building descriptors fails because the source DescriptorProtos
-   * are not valid.
-   */
-  public static class DescriptorValidationException extends Exception {
-    private static final long serialVersionUID = 5750205775490483148L;
-
-    /** Gets the full name of the descriptor where the error occurred. */
-    public String getProblemSymbolName() { return name; }
-
-    /**
-     * Gets the protocol message representation of the invalid descriptor.
-     */
-    public Message getProblemProto() { return proto; }
-
-    /**
-     * Gets a human-readable description of the error.
-     */
-    public String getDescription() { return description; }
-
-    private final String name;
-    private final Message proto;
-    private final String description;
-
-    private DescriptorValidationException(
-        final GenericDescriptor problemDescriptor,
-        final String description) {
-      super(problemDescriptor.getFullName() + ": " + description);
-
-      // Note that problemDescriptor may be partially uninitialized, so we
-      // don't want to expose it directly to the user.  So, we only provide
-      // the name and the original proto.
-      name = problemDescriptor.getFullName();
-      proto = problemDescriptor.toProto();
-      this.description = description;
-    }
-
-    private DescriptorValidationException(
-        final GenericDescriptor problemDescriptor,
-        final String description,
-        final Throwable cause) {
-      this(problemDescriptor, description);
-      initCause(cause);
-    }
-
-    private DescriptorValidationException(
-        final FileDescriptor problemDescriptor,
-        final String description) {
-      super(problemDescriptor.getName() + ": " + description);
-
-      // Note that problemDescriptor may be partially uninitialized, so we
-      // don't want to expose it directly to the user.  So, we only provide
-      // the name and the original proto.
-      name = problemDescriptor.getName();
-      proto = problemDescriptor.toProto();
-      this.description = description;
-    }
-  }
-
-  // =================================================================
-
-  /**
-   * A private helper class which contains lookup tables containing all the
-   * descriptors defined in a particular file.
-   */
-  private static final class DescriptorPool {
-
-    /** Defines what subclass of descriptors to search in the descriptor pool.
-     */
-    enum SearchFilter {
-      TYPES_ONLY, AGGREGATES_ONLY, ALL_SYMBOLS
-    }
-
-    DescriptorPool(final FileDescriptor[] dependencies,
-        boolean allowUnknownDependencies) {
-      this.dependencies = new HashSet<FileDescriptor>();
-      this.allowUnknownDependencies = allowUnknownDependencies;
-
-      for (int i = 0; i < dependencies.length; i++) {
-        this.dependencies.add(dependencies[i]);
-        importPublicDependencies(dependencies[i]);
-      }
-
-      for (final FileDescriptor dependency : this.dependencies) {
-        try {
-          addPackage(dependency.getPackage(), dependency);
-        } catch (DescriptorValidationException e) {
-          // Can't happen, because addPackage() only fails when the name
-          // conflicts with a non-package, but we have not yet added any
-          // non-packages at this point.
-          throw new AssertionError(e);
-        }
-      }
-    }
-
-    /** Find and put public dependencies of the file into dependencies set.*/
-    private void importPublicDependencies(final FileDescriptor file) {
-      for (FileDescriptor dependency : file.getPublicDependencies()) {
-        if (dependencies.add(dependency)) {
-          importPublicDependencies(dependency);
-        }
-      }
-    }
-
-    private final Set<FileDescriptor> dependencies;
-    private boolean allowUnknownDependencies;
-
-    private final Map<String, GenericDescriptor> descriptorsByName =
-      new HashMap<String, GenericDescriptor>();
-    private final Map<DescriptorIntPair, FieldDescriptor> fieldsByNumber =
-      new HashMap<DescriptorIntPair, FieldDescriptor>();
-    private final Map<DescriptorIntPair, EnumValueDescriptor> enumValuesByNumber
-        = new HashMap<DescriptorIntPair, EnumValueDescriptor>();
-
-    /** Find a generic descriptor by fully-qualified name. */
-    GenericDescriptor findSymbol(final String fullName) {
-      return findSymbol(fullName, SearchFilter.ALL_SYMBOLS);
-    }
-
-    /** Find a descriptor by fully-qualified name and given option to only
-     * search valid field type descriptors.
-     */
-    GenericDescriptor findSymbol(final String fullName,
-                                 final SearchFilter filter) {
-      GenericDescriptor result = descriptorsByName.get(fullName);
-      if (result != null) {
-        if ((filter==SearchFilter.ALL_SYMBOLS) ||
-            ((filter==SearchFilter.TYPES_ONLY) && isType(result)) ||
-            ((filter==SearchFilter.AGGREGATES_ONLY) && isAggregate(result))) {
-          return result;
-        }
-      }
-
-      for (final FileDescriptor dependency : dependencies) {
-        result = dependency.pool.descriptorsByName.get(fullName);
-        if (result != null) {
-          if ((filter==SearchFilter.ALL_SYMBOLS) ||
-              ((filter==SearchFilter.TYPES_ONLY) && isType(result)) ||
-              ((filter==SearchFilter.AGGREGATES_ONLY) && isAggregate(result))) {
-            return result;
-          }
-        }
-      }
-
-      return null;
-    }
-
-    /** Checks if the descriptor is a valid type for a message field. */
-    boolean isType(GenericDescriptor descriptor) {
-      return (descriptor instanceof Descriptor) ||
-        (descriptor instanceof EnumDescriptor);
-    }
-
-    /** Checks if the descriptor is a valid namespace type. */
-    boolean isAggregate(GenericDescriptor descriptor) {
-      return (descriptor instanceof Descriptor) ||
-        (descriptor instanceof EnumDescriptor) ||
-        (descriptor instanceof PackageDescriptor) ||
-        (descriptor instanceof ServiceDescriptor);
-    }
-
-    /**
-     * Look up a type descriptor by name, relative to some other descriptor.
-     * The name may be fully-qualified (with a leading '.'),
-     * partially-qualified, or unqualified.  C++-like name lookup semantics
-     * are used to search for the matching descriptor.
-     */
-    GenericDescriptor lookupSymbol(final String name,
-                                   final GenericDescriptor relativeTo,
-                                   final DescriptorPool.SearchFilter filter)
-                            throws DescriptorValidationException {
-      // TODO(kenton):  This could be optimized in a number of ways.
-
-      GenericDescriptor result;
-      String fullname;
-      if (name.startsWith(".")) {
-        // Fully-qualified name.
-        fullname = name.substring(1);
-        result = findSymbol(fullname, filter);
-      } else {
-        // If "name" is a compound identifier, we want to search for the
-        // first component of it, then search within it for the rest.
-        // If name is something like "Foo.Bar.baz", and symbols named "Foo" are
-        // defined in multiple parent scopes, we only want to find "Bar.baz" in
-        // the innermost one.  E.g., the following should produce an error:
-        //   message Bar { message Baz {} }
-        //   message Foo {
-        //     message Bar {
-        //     }
-        //     optional Bar.Baz baz = 1;
-        //   }
-        // So, we look for just "Foo" first, then look for "Bar.baz" within it
-        // if found.
-        final int firstPartLength = name.indexOf('.');
-        final String firstPart;
-        if (firstPartLength == -1) {
-          firstPart = name;
-        } else {
-          firstPart = name.substring(0, firstPartLength);
-        }
-
-        // We will search each parent scope of "relativeTo" looking for the
-        // symbol.
-        final StringBuilder scopeToTry =
-            new StringBuilder(relativeTo.getFullName());
-
-        while (true) {
-          // Chop off the last component of the scope.
-          final int dotpos = scopeToTry.lastIndexOf(".");
-          if (dotpos == -1) {
-            fullname = name;
-            result = findSymbol(name, filter);
-            break;
-          } else {
-            scopeToTry.setLength(dotpos + 1);
-
-            // Append firstPart and try to find
-            scopeToTry.append(firstPart);
-            result = findSymbol(scopeToTry.toString(),
-                DescriptorPool.SearchFilter.AGGREGATES_ONLY);
-
-            if (result != null) {
-              if (firstPartLength != -1) {
-                // We only found the first part of the symbol.  Now look for
-                // the whole thing.  If this fails, we *don't* want to keep
-                // searching parent scopes.
-                scopeToTry.setLength(dotpos + 1);
-                scopeToTry.append(name);
-                result = findSymbol(scopeToTry.toString(), filter);
-              }
-              fullname = scopeToTry.toString();
-              break;
-            }
-
-            // Not found.  Remove the name so we can try again.
-            scopeToTry.setLength(dotpos);
-          }
-        }
-      }
-
-      if (result == null) {
-        if (allowUnknownDependencies && filter == SearchFilter.TYPES_ONLY) {
-          logger.warning("The descriptor for message type \"" + name +
-              "\" can not be found and a placeholder is created for it");
-          // We create a dummy message descriptor here regardless of the
-          // expected type. If the type should be message, this dummy
-          // descriptor will work well and if the type should be enum, a
-          // DescriptorValidationException will be thrown latter. In either
-          // case, the code works as expected: we allow unknown message types
-          // but not unknwon enum types.
-          result = new Descriptor(fullname);
-          // Add the placeholder file as a dependency so we can find the
-          // placeholder symbol when resolving other references.
-          this.dependencies.add(result.getFile());
-          return result;
-        } else {
-          throw new DescriptorValidationException(relativeTo,
-              '\"' + name + "\" is not defined.");
-        }
-      } else {
-        return result;
-      }
-    }
-
-    /**
-     * Adds a symbol to the symbol table.  If a symbol with the same name
-     * already exists, throws an error.
-     */
-    void addSymbol(final GenericDescriptor descriptor)
-            throws DescriptorValidationException {
-      validateSymbolName(descriptor);
-
-      final String fullName = descriptor.getFullName();
-      final int dotpos = fullName.lastIndexOf('.');
-
-      final GenericDescriptor old = descriptorsByName.put(fullName, descriptor);
-      if (old != null) {
-        descriptorsByName.put(fullName, old);
-
-        if (descriptor.getFile() == old.getFile()) {
-          if (dotpos == -1) {
-            throw new DescriptorValidationException(descriptor,
-                '\"' + fullName + "\" is already defined.");
-          } else {
-            throw new DescriptorValidationException(descriptor,
-                '\"' + fullName.substring(dotpos + 1) +
-              "\" is already defined in \"" +
-              fullName.substring(0, dotpos) + "\".");
-          }
-        } else {
-          throw new DescriptorValidationException(descriptor,
-              '\"' + fullName + "\" is already defined in file \"" +
-            old.getFile().getName() + "\".");
-        }
-      }
-    }
-
-    /**
-     * Represents a package in the symbol table.  We use PackageDescriptors
-     * just as placeholders so that someone cannot define, say, a message type
-     * that has the same name as an existing package.
-     */
-    private static final class PackageDescriptor extends GenericDescriptor {
-      @Override
-      public Message toProto() {
-        return file.toProto();
-      }
-      @Override
-      public String getName() {
-        return name;
-      }
-      @Override
-      public String getFullName() {
-        return fullName;
-      }
-      @Override
-      public FileDescriptor getFile() {
-        return file;
-      }
-
-      PackageDescriptor(final String name, final String fullName,
-                        final FileDescriptor file) {
-        this.file = file;
-        this.fullName = fullName;
-        this.name = name;
-      }
-
-      private final String name;
-      private final String fullName;
-      private final FileDescriptor file;
-    }
-
-    /**
-     * Adds a package to the symbol tables.  If a package by the same name
-     * already exists, that is fine, but if some other kind of symbol exists
-     * under the same name, an exception is thrown.  If the package has
-     * multiple components, this also adds the parent package(s).
-     */
-    void addPackage(final String fullName, final FileDescriptor file)
-             throws DescriptorValidationException {
-      final int dotpos = fullName.lastIndexOf('.');
-      final String name;
-      if (dotpos == -1) {
-        name = fullName;
-      } else {
-        addPackage(fullName.substring(0, dotpos), file);
-        name = fullName.substring(dotpos + 1);
-      }
-
-      final GenericDescriptor old =
-        descriptorsByName.put(fullName,
-          new PackageDescriptor(name, fullName, file));
-      if (old != null) {
-        descriptorsByName.put(fullName, old);
-        if (!(old instanceof PackageDescriptor)) {
-          throw new DescriptorValidationException(file,
-              '\"' + name + "\" is already defined (as something other than a "
-              + "package) in file \"" + old.getFile().getName() + "\".");
-        }
-      }
-    }
-
-    /** A (GenericDescriptor, int) pair, used as a map key. */
-    private static final class DescriptorIntPair {
-      private final GenericDescriptor descriptor;
-      private final int number;
-
-      DescriptorIntPair(final GenericDescriptor descriptor, final int number) {
-        this.descriptor = descriptor;
-        this.number = number;
-      }
-
-      @Override
-      public int hashCode() {
-        return descriptor.hashCode() * ((1 << 16) - 1) + number;
-      }
-      @Override
-      public boolean equals(final Object obj) {
-        if (!(obj instanceof DescriptorIntPair)) {
-          return false;
-        }
-        final DescriptorIntPair other = (DescriptorIntPair)obj;
-        return descriptor == other.descriptor && number == other.number;
-      }
-    }
-
-    /**
-     * Adds a field to the fieldsByNumber table.  Throws an exception if a
-     * field with the same containing type and number already exists.
-     */
-    void addFieldByNumber(final FieldDescri

<TRUNCATED>

[16/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcController.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcController.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcController.java
deleted file mode 100644
index 99e306b..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcController.java
+++ /dev/null
@@ -1,118 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <p>An {@code RpcController} mediates a single method call.  The primary
- * purpose of the controller is to provide a way to manipulate settings
- * specific to the RPC implementation and to find out about RPC-level errors.
- *
- * <p>Starting with version 2.3.0, RPC implementations should not try to build
- * on this, but should instead provide code generator plugins which generate
- * code specific to the particular RPC implementation.  This way the generated
- * code can be more appropriate for the implementation in use and can avoid
- * unnecessary layers of indirection.
- *
- * <p>The methods provided by the {@code RpcController} interface are intended
- * to be a "least common denominator" set of features which we expect all
- * implementations to support.  Specific implementations may provide more
- * advanced features (e.g. deadline propagation).
- *
- * @author kenton@google.com Kenton Varda
- */
-public interface RpcController {
-  // -----------------------------------------------------------------
-  // These calls may be made from the client side only.  Their results
-  // are undefined on the server side (may throw RuntimeExceptions).
-
-  /**
-   * Resets the RpcController to its initial state so that it may be reused in
-   * a new call.  This can be called from the client side only.  It must not
-   * be called while an RPC is in progress.
-   */
-  void reset();
-
-  /**
-   * After a call has finished, returns true if the call failed.  The possible
-   * reasons for failure depend on the RPC implementation.  {@code failed()}
-   * most only be called on the client side, and must not be called before a
-   * call has finished.
-   */
-  boolean failed();
-
-  /**
-   * If {@code failed()} is {@code true}, returns a human-readable description
-   * of the error.
-   */
-  String errorText();
-
-  /**
-   * Advises the RPC system that the caller desires that the RPC call be
-   * canceled.  The RPC system may cancel it immediately, may wait awhile and
-   * then cancel it, or may not even cancel the call at all.  If the call is
-   * canceled, the "done" callback will still be called and the RpcController
-   * will indicate that the call failed at that time.
-   */
-  void startCancel();
-
-  // -----------------------------------------------------------------
-  // These calls may be made from the server side only.  Their results
-  // are undefined on the client side (may throw RuntimeExceptions).
-
-  /**
-   * Causes {@code failed()} to return true on the client side.  {@code reason}
-   * will be incorporated into the message returned by {@code errorText()}.
-   * If you find you need to return machine-readable information about
-   * failures, you should incorporate it into your response protocol buffer
-   * and should NOT call {@code setFailed()}.
-   */
-  void setFailed(String reason);
-
-  /**
-   * If {@code true}, indicates that the client canceled the RPC, so the server
-   * may as well give up on replying to it.  This method must be called on the
-   * server side only.  The server should still call the final "done" callback.
-   */
-  boolean isCanceled();
-
-  /**
-   * Asks that the given callback be called when the RPC is canceled.  The
-   * parameter passed to the callback will always be {@code null}.  The
-   * callback will always be called exactly once.  If the RPC completes without
-   * being canceled, the callback will be called after completion.  If the RPC
-   * has already been canceled when NotifyOnCancel() is called, the callback
-   * will be called immediately.
-   *
-   * <p>{@code notifyOnCancel()} must be called no more than once per request.
-   * It must be called on the server side only.
-   */
-  void notifyOnCancel(RpcCallback<Object> callback);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcUtil.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcUtil.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcUtil.java
deleted file mode 100644
index c2a6878..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/RpcUtil.java
+++ /dev/null
@@ -1,136 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Grab-bag of utility functions useful when dealing with RPCs.
- *
- * @author kenton@google.com Kenton Varda
- */
-public final class RpcUtil {
-  private RpcUtil() {}
-
-  /**
-   * Take an {@code RpcCallback<Message>} and convert it to an
-   * {@code RpcCallback} accepting a specific message type.  This is always
-   * type-safe (parameter type contravariance).
-   */
-  @SuppressWarnings("unchecked")
-  public static <Type extends Message> RpcCallback<Type>
-  specializeCallback(final RpcCallback<Message> originalCallback) {
-    return (RpcCallback<Type>)originalCallback;
-    // The above cast works, but only due to technical details of the Java
-    // implementation.  A more theoretically correct -- but less efficient --
-    // implementation would be as follows:
-    //   return new RpcCallback<Type>() {
-    //     public void run(Type parameter) {
-    //       originalCallback.run(parameter);
-    //     }
-    //   };
-  }
-
-  /**
-   * Take an {@code RpcCallback} accepting a specific message type and convert
-   * it to an {@code RpcCallback<Message>}.  The generalized callback will
-   * accept any message object which has the same descriptor, and will convert
-   * it to the correct class before calling the original callback.  However,
-   * if the generalized callback is given a message with a different descriptor,
-   * an exception will be thrown.
-   */
-  public static <Type extends Message>
-  RpcCallback<Message> generalizeCallback(
-      final RpcCallback<Type> originalCallback,
-      final Class<Type> originalClass,
-      final Type defaultInstance) {
-    return new RpcCallback<Message>() {
-      @Override
-      public void run(final Message parameter) {
-        Type typedParameter;
-        try {
-          typedParameter = originalClass.cast(parameter);
-        } catch (ClassCastException ignored) {
-          typedParameter = copyAsType(defaultInstance, parameter);
-        }
-        originalCallback.run(typedParameter);
-      }
-    };
-  }
-
-  /**
-   * Creates a new message of type "Type" which is a copy of "source".  "source"
-   * must have the same descriptor but may be a different class (e.g.
-   * DynamicMessage).
-   */
-  @SuppressWarnings("unchecked")
-  private static <Type extends Message> Type copyAsType(
-      final Type typeDefaultInstance, final Message source) {
-    return (Type) typeDefaultInstance
-        .newBuilderForType().mergeFrom(source).build();
-  }
-
-  /**
-   * Creates a callback which can only be called once.  This may be useful for
-   * security, when passing a callback to untrusted code:  most callbacks do
-   * not expect to be called more than once, so doing so may expose bugs if it
-   * is not prevented.
-   */
-  public static <ParameterType>
-    RpcCallback<ParameterType> newOneTimeCallback(
-      final RpcCallback<ParameterType> originalCallback) {
-    return new RpcCallback<ParameterType>() {
-      private boolean alreadyCalled = false;
-
-      @Override
-      public void run(final ParameterType parameter) {
-        synchronized (this) {
-          if (alreadyCalled) {
-            throw new AlreadyCalledException();
-          }
-          alreadyCalled = true;
-        }
-
-        originalCallback.run(parameter);
-      }
-    };
-  }
-
-  /**
-   * Exception thrown when a one-time callback is called more than once.
-   */
-  public static final class AlreadyCalledException extends RuntimeException {
-    private static final long serialVersionUID = 5469741279507848266L;
-
-    public AlreadyCalledException() {
-      super("This RpcCallback was already called and cannot be called " +
-            "multiple times.");
-    }
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Service.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Service.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Service.java
deleted file mode 100644
index 849d9ff..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Service.java
+++ /dev/null
@@ -1,117 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Abstract base interface for protocol-buffer-based RPC services.  Services
- * themselves are abstract classes (implemented either by servers or as
- * stubs), but they subclass this base interface.  The methods of this
- * interface can be used to call the methods of the service without knowing
- * its exact type at compile time (analogous to the Message interface).
- *
- * <p>Starting with version 2.3.0, RPC implementations should not try to build
- * on this, but should instead provide code generator plugins which generate
- * code specific to the particular RPC implementation.  This way the generated
- * code can be more appropriate for the implementation in use and can avoid
- * unnecessary layers of indirection.
- *
- * @author kenton@google.com Kenton Varda
- */
-public interface Service {
-  /**
-   * Get the {@code ServiceDescriptor} describing this service and its methods.
-   */
-  Descriptors.ServiceDescriptor getDescriptorForType();
-
-  /**
-   * <p>Call a method of the service specified by MethodDescriptor.  This is
-   * normally implemented as a simple {@code switch()} that calls the standard
-   * definitions of the service's methods.
-   *
-   * <p>Preconditions:
-   * <ul>
-   *   <li>{@code method.getService() == getDescriptorForType()}
-   *   <li>{@code request} is of the exact same class as the object returned by
-   *       {@code getRequestPrototype(method)}.
-   *   <li>{@code controller} is of the correct type for the RPC implementation
-   *       being used by this Service.  For stubs, the "correct type" depends
-   *       on the RpcChannel which the stub is using.  Server-side Service
-   *       implementations are expected to accept whatever type of
-   *       {@code RpcController} the server-side RPC implementation uses.
-   * </ul>
-   *
-   * <p>Postconditions:
-   * <ul>
-   *   <li>{@code done} will be called when the method is complete.  This may be
-   *       before {@code callMethod()} returns or it may be at some point in
-   *       the future.
-   *   <li>The parameter to {@code done} is the response.  It must be of the
-   *       exact same type as would be returned by
-   *       {@code getResponsePrototype(method)}.
-   *   <li>If the RPC failed, the parameter to {@code done} will be
-   *       {@code null}.  Further details about the failure can be found by
-   *       querying {@code controller}.
-   * </ul>
-   */
-  void callMethod(Descriptors.MethodDescriptor method,
-                  RpcController controller,
-                  Message request,
-                  RpcCallback<Message> done);
-
-  /**
-   * <p>{@code callMethod()} requires that the request passed in is of a
-   * particular subclass of {@code Message}.  {@code getRequestPrototype()}
-   * gets the default instances of this type for a given method.  You can then
-   * call {@code Message.newBuilderForType()} on this instance to
-   * construct a builder to build an object which you can then pass to
-   * {@code callMethod()}.
-   *
-   * <p>Example:
-   * <pre>
-   *   MethodDescriptor method =
-   *     service.getDescriptorForType().findMethodByName("Foo");
-   *   Message request =
-   *     stub.getRequestPrototype(method).newBuilderForType()
-   *         .mergeFrom(input).build();
-   *   service.callMethod(method, request, callback);
-   * </pre>
-   */
-  Message getRequestPrototype(Descriptors.MethodDescriptor method);
-
-  /**
-   * Like {@code getRequestPrototype()}, but gets a prototype of the response
-   * message.  {@code getResponsePrototype()} is generally not needed because
-   * the {@code Service} implementation constructs the response message itself,
-   * but it may be useful in some cases to know ahead of time what type of
-   * object will be returned.
-   */
-  Message getResponsePrototype(Descriptors.MethodDescriptor method);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ServiceException.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ServiceException.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ServiceException.java
deleted file mode 100644
index 9b04a12..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ServiceException.java
+++ /dev/null
@@ -1,52 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * Thrown by blocking RPC methods when a failure occurs.
- *
- * @author cpovirk@google.com (Chris Povirk)
- */
-public class ServiceException extends Exception {
-  private static final long serialVersionUID = -1219262335729891920L;
-
-  public ServiceException(final String message) {
-    super(message);
-  }
-
-  public ServiceException(final Throwable cause) {
-    super(cause);
-  }
-
-  public ServiceException(final String message, final Throwable cause) {
-    super(message, cause);
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SingleFieldBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SingleFieldBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SingleFieldBuilder.java
deleted file mode 100644
index 818901a..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SingleFieldBuilder.java
+++ /dev/null
@@ -1,241 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * {@code SingleFieldBuilder} implements a structure that a protocol
- * message uses to hold a single field of another protocol message. It supports
- * the classical use case of setting an immutable {@link Message} as the value
- * of the field and is highly optimized around this.
- * <br>
- * It also supports the additional use case of setting a {@link Message.Builder}
- * as the field and deferring conversion of that {@code Builder}
- * to an immutable {@code Message}. In this way, it's possible to maintain
- * a tree of {@code Builder}'s that acts as a fully read/write data
- * structure.
- * <br>
- * Logically, one can think of a tree of builders as converting the entire tree
- * to messages when build is called on the root or when any method is called
- * that desires a Message instead of a Builder. In terms of the implementation,
- * the {@code SingleFieldBuilder} and {@code RepeatedFieldBuilder}
- * classes cache messages that were created so that messages only need to be
- * created when some change occurred in its builder or a builder for one of its
- * descendants.
- *
- * @param <MType> the type of message for the field
- * @param <BType> the type of builder for the field
- * @param <IType> the common interface for the message and the builder
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public class SingleFieldBuilder
-    <MType extends GeneratedMessage,
-     BType extends GeneratedMessage.Builder,
-     IType extends MessageOrBuilder>
-    implements GeneratedMessage.BuilderParent {
-
-  // Parent to send changes to.
-  private GeneratedMessage.BuilderParent parent;
-
-  // Invariant: one of builder or message fields must be non-null.
-
-  // If set, this is the case where we are backed by a builder. In this case,
-  // message field represents a cached message for the builder (or null if
-  // there is no cached message).
-  private BType builder;
-
-  // If builder is non-null, this represents a cached message from the builder.
-  // If builder is null, this is the authoritative message for the field.
-  private MType message;
-
-  // Indicates that we've built a message and so we are now obligated
-  // to dispatch dirty invalidations. See GeneratedMessage.BuilderListener.
-  private boolean isClean;
-
-  public SingleFieldBuilder(
-      MType message,
-      GeneratedMessage.BuilderParent parent,
-      boolean isClean) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    this.message = message;
-    this.parent = parent;
-    this.isClean = isClean;
-  }
-
-  public void dispose() {
-    // Null out parent so we stop sending it invalidations.
-    parent = null;
-  }
-
-  /**
-   * Get the message for the field. If the message is currently stored
-   * as a {@code Builder}, it is converted to a {@code Message} by
-   * calling {@link Message.Builder#buildPartial} on it. If no message has
-   * been set, returns the default instance of the message.
-   *
-   * @return the message for the field
-   */
-  @SuppressWarnings("unchecked")
-  public MType getMessage() {
-    if (message == null) {
-      // If message is null, the invariant is that we must be have a builder.
-      message = (MType) builder.buildPartial();
-    }
-    return message;
-  }
-
-  /**
-   * Builds the message and returns it.
-   *
-   * @return the message
-   */
-  public MType build() {
-    // Now that build has been called, we are required to dispatch
-    // invalidations.
-    isClean = true;
-    return getMessage();
-  }
-
-  /**
-   * Gets a builder for the field. If no builder has been created yet, a
-   * builder is created on demand by calling {@link Message#toBuilder}.
-   *
-   * @return The builder for the field
-   */
-  @SuppressWarnings("unchecked")
-  public BType getBuilder() {
-    if (builder == null) {
-      // builder.mergeFrom() on a fresh builder
-      // does not create any sub-objects with independent clean/dirty states,
-      // therefore setting the builder itself to clean without actually calling
-      // build() cannot break any invariants.
-      builder = (BType) message.newBuilderForType(this);
-      builder.mergeFrom(message); // no-op if message is the default message
-      builder.markClean();
-    }
-    return builder;
-  }
-
-  /**
-   * Gets the base class interface for the field. This may either be a builder
-   * or a message. It will return whatever is more efficient.
-   *
-   * @return the message or builder for the field as the base class interface
-   */
-  @SuppressWarnings("unchecked")
-  public IType getMessageOrBuilder() {
-    if (builder != null) {
-      return  (IType) builder;
-    } else {
-      return (IType) message;
-    }
-  }
-
-  /**
-   * Sets a  message for the field replacing any existing value.
-   *
-   * @param message the message to set
-   * @return the builder
-   */
-  public SingleFieldBuilder<MType, BType, IType> setMessage(
-      MType message) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    this.message = message;
-    if (builder != null) {
-      builder.dispose();
-      builder = null;
-    }
-    onChanged();
-    return this;
-  }
-
-  /**
-   * Merges the field from another field.
-   *
-   * @param value the value to merge from
-   * @return the builder
-   */
-  public SingleFieldBuilder<MType, BType, IType> mergeFrom(
-      MType value) {
-    if (builder == null && message == message.getDefaultInstanceForType()) {
-      message = value;
-    } else {
-      getBuilder().mergeFrom(value);
-    }
-    onChanged();
-    return this;
-  }
-
-  /**
-   * Clears the value of the field.
-   *
-   * @return the builder
-   */
-  @SuppressWarnings("unchecked")
-  public SingleFieldBuilder<MType, BType, IType> clear() {
-    message = (MType) (message != null ?
-        message.getDefaultInstanceForType() :
-        builder.getDefaultInstanceForType());
-    if (builder != null) {
-      builder.dispose();
-      builder = null;
-    }
-    onChanged();
-    return this;
-  }
-
-  /**
-   * Called when a the builder or one of its nested children has changed
-   * and any parent should be notified of its invalidation.
-   */
-  private void onChanged() {
-    // If builder is null, this is the case where onChanged is being called
-    // from setMessage or clear.
-    if (builder != null) {
-      message = null;
-    }
-    if (isClean && parent != null) {
-      parent.markDirty();
-
-      // Don't keep dispatching invalidations until build is called again.
-      isClean = false;
-    }
-  }
-
-  @Override
-  public void markDirty() {
-    onChanged();
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SingleFieldBuilderV3.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SingleFieldBuilderV3.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SingleFieldBuilderV3.java
deleted file mode 100644
index 0a294dc..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SingleFieldBuilderV3.java
+++ /dev/null
@@ -1,241 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * {@code SingleFieldBuilderV3} implements a structure that a protocol
- * message uses to hold a single field of another protocol message. It supports
- * the classical use case of setting an immutable {@link Message} as the value
- * of the field and is highly optimized around this.
- * <br>
- * It also supports the additional use case of setting a {@link Message.Builder}
- * as the field and deferring conversion of that {@code Builder}
- * to an immutable {@code Message}. In this way, it's possible to maintain
- * a tree of {@code Builder}'s that acts as a fully read/write data
- * structure.
- * <br>
- * Logically, one can think of a tree of builders as converting the entire tree
- * to messages when build is called on the root or when any method is called
- * that desires a Message instead of a Builder. In terms of the implementation,
- * the {@code SingleFieldBuilderV3} and {@code RepeatedFieldBuilderV3}
- * classes cache messages that were created so that messages only need to be
- * created when some change occurred in its builder or a builder for one of its
- * descendants.
- *
- * @param <MType> the type of message for the field
- * @param <BType> the type of builder for the field
- * @param <IType> the common interface for the message and the builder
- *
- * @author jonp@google.com (Jon Perlow)
- */
-public class SingleFieldBuilderV3
-    <MType extends AbstractMessage,
-     BType extends AbstractMessage.Builder,
-     IType extends MessageOrBuilder>
-    implements AbstractMessage.BuilderParent {
-
-  // Parent to send changes to.
-  private AbstractMessage.BuilderParent parent;
-
-  // Invariant: one of builder or message fields must be non-null.
-
-  // If set, this is the case where we are backed by a builder. In this case,
-  // message field represents a cached message for the builder (or null if
-  // there is no cached message).
-  private BType builder;
-
-  // If builder is non-null, this represents a cached message from the builder.
-  // If builder is null, this is the authoritative message for the field.
-  private MType message;
-
-  // Indicates that we've built a message and so we are now obligated
-  // to dispatch dirty invalidations. See AbstractMessage.BuilderListener.
-  private boolean isClean;
-
-  public SingleFieldBuilderV3(
-      MType message,
-      AbstractMessage.BuilderParent parent,
-      boolean isClean) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    this.message = message;
-    this.parent = parent;
-    this.isClean = isClean;
-  }
-
-  public void dispose() {
-    // Null out parent so we stop sending it invalidations.
-    parent = null;
-  }
-
-  /**
-   * Get the message for the field. If the message is currently stored
-   * as a {@code Builder}, it is converted to a {@code Message} by
-   * calling {@link Message.Builder#buildPartial} on it. If no message has
-   * been set, returns the default instance of the message.
-   *
-   * @return the message for the field
-   */
-  @SuppressWarnings("unchecked")
-  public MType getMessage() {
-    if (message == null) {
-      // If message is null, the invariant is that we must be have a builder.
-      message = (MType) builder.buildPartial();
-    }
-    return message;
-  }
-
-  /**
-   * Builds the message and returns it.
-   *
-   * @return the message
-   */
-  public MType build() {
-    // Now that build has been called, we are required to dispatch
-    // invalidations.
-    isClean = true;
-    return getMessage();
-  }
-
-  /**
-   * Gets a builder for the field. If no builder has been created yet, a
-   * builder is created on demand by calling {@link Message#toBuilder}.
-   *
-   * @return The builder for the field
-   */
-  @SuppressWarnings("unchecked")
-  public BType getBuilder() {
-    if (builder == null) {
-      // builder.mergeFrom() on a fresh builder
-      // does not create any sub-objects with independent clean/dirty states,
-      // therefore setting the builder itself to clean without actually calling
-      // build() cannot break any invariants.
-      builder = (BType) message.newBuilderForType(this);
-      builder.mergeFrom(message); // no-op if message is the default message
-      builder.markClean();
-    }
-    return builder;
-  }
-
-  /**
-   * Gets the base class interface for the field. This may either be a builder
-   * or a message. It will return whatever is more efficient.
-   *
-   * @return the message or builder for the field as the base class interface
-   */
-  @SuppressWarnings("unchecked")
-  public IType getMessageOrBuilder() {
-    if (builder != null) {
-      return  (IType) builder;
-    } else {
-      return (IType) message;
-    }
-  }
-
-  /**
-   * Sets a  message for the field replacing any existing value.
-   *
-   * @param message the message to set
-   * @return the builder
-   */
-  public SingleFieldBuilderV3<MType, BType, IType> setMessage(
-      MType message) {
-    if (message == null) {
-      throw new NullPointerException();
-    }
-    this.message = message;
-    if (builder != null) {
-      builder.dispose();
-      builder = null;
-    }
-    onChanged();
-    return this;
-  }
-
-  /**
-   * Merges the field from another field.
-   *
-   * @param value the value to merge from
-   * @return the builder
-   */
-  public SingleFieldBuilderV3<MType, BType, IType> mergeFrom(
-      MType value) {
-    if (builder == null && message == message.getDefaultInstanceForType()) {
-      message = value;
-    } else {
-      getBuilder().mergeFrom(value);
-    }
-    onChanged();
-    return this;
-  }
-
-  /**
-   * Clears the value of the field.
-   *
-   * @return the builder
-   */
-  @SuppressWarnings("unchecked")
-  public SingleFieldBuilderV3<MType, BType, IType> clear() {
-    message = (MType) (message != null ?
-        message.getDefaultInstanceForType() :
-        builder.getDefaultInstanceForType());
-    if (builder != null) {
-      builder.dispose();
-      builder = null;
-    }
-    onChanged();
-    return this;
-  }
-
-  /**
-   * Called when a the builder or one of its nested children has changed
-   * and any parent should be notified of its invalidation.
-   */
-  private void onChanged() {
-    // If builder is null, this is the case where onChanged is being called
-    // from setMessage or clear.
-    if (builder != null) {
-      message = null;
-    }
-    if (isClean && parent != null) {
-      parent.markDirty();
-
-      // Don't keep dispatching invalidations until build is called again.
-      isClean = false;
-    }
-  }
-
-  @Override
-  public void markDirty() {
-    onChanged();
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SmallSortedMap.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SmallSortedMap.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SmallSortedMap.java
deleted file mode 100644
index a24c2ce..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SmallSortedMap.java
+++ /dev/null
@@ -1,673 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.AbstractMap;
-import java.util.AbstractSet;
-import java.util.ArrayList;
-import java.util.Collections;
-import java.util.Iterator;
-import java.util.List;
-import java.util.Map;
-import java.util.NoSuchElementException;
-import java.util.Set;
-import java.util.SortedMap;
-import java.util.TreeMap;
-
-/**
- * A custom map implementation from FieldDescriptor to Object optimized to
- * minimize the number of memory allocations for instances with a small number
- * of mappings. The implementation stores the first {@code k} mappings in an
- * array for a configurable value of {@code k}, allowing direct access to the
- * corresponding {@code Entry}s without the need to create an Iterator. The
- * remaining entries are stored in an overflow map. Iteration over the entries
- * in the map should be done as follows:
- *
- * <pre>   {@code
- * for (int i = 0; i < fieldMap.getNumArrayEntries(); i++) {
- *   process(fieldMap.getArrayEntryAt(i));
- * }
- * for (Map.Entry<K, V> entry : fieldMap.getOverflowEntries()) {
- *   process(entry);
- * }
- * }</pre>
- *
- * The resulting iteration is in order of ascending field tag number. The
- * object returned by {@link #entrySet()} adheres to the same contract but is
- * less efficient as it necessarily involves creating an object for iteration.
- * <p>
- * The tradeoff for this memory efficiency is that the worst case running time
- * of the {@code put()} operation is {@code O(k + lg n)}, which happens when
- * entries are added in descending order. {@code k} should be chosen such that
- * it covers enough common cases without adversely affecting larger maps. In
- * practice, the worst case scenario does not happen for extensions because
- * extension fields are serialized and deserialized in order of ascending tag
- * number, but the worst case scenario can happen for DynamicMessages.
- * <p>
- * The running time for all other operations is similar to that of
- * {@code TreeMap}.
- * <p>
- * Instances are not thread-safe until {@link #makeImmutable()} is called,
- * after which any modifying operation will result in an
- * {@link UnsupportedOperationException}.
- *
- * @author darick@google.com Darick Tong
- */
-// This class is final for all intents and purposes because the constructor is
-// private. However, the FieldDescriptor-specific logic is encapsulated in
-// a subclass to aid testability of the core logic.
-class SmallSortedMap<K extends Comparable<K>, V> extends AbstractMap<K, V> {
-
-  /**
-   * Creates a new instance for mapping FieldDescriptors to their values.
-   * The {@link #makeImmutable()} implementation will convert the List values
-   * of any repeated fields to unmodifiable lists.
-   *
-   * @param arraySize The size of the entry array containing the
-   *        lexicographically smallest mappings.
-   */
-  static <FieldDescriptorType extends
-      FieldSet.FieldDescriptorLite<FieldDescriptorType>>
-      SmallSortedMap<FieldDescriptorType, Object> newFieldMap(int arraySize) {
-    return new SmallSortedMap<FieldDescriptorType, Object>(arraySize) {
-      @Override
-      @SuppressWarnings("unchecked")
-      public void makeImmutable() {
-        if (!isImmutable()) {
-          for (int i = 0; i < getNumArrayEntries(); i++) {
-            final Map.Entry<FieldDescriptorType, Object> entry =
-                getArrayEntryAt(i);
-            if (entry.getKey().isRepeated()) {
-              final List value = (List) entry.getValue();
-              entry.setValue(Collections.unmodifiableList(value));
-            }
-          }
-          for (Map.Entry<FieldDescriptorType, Object> entry :
-                   getOverflowEntries()) {
-            if (entry.getKey().isRepeated()) {
-              final List value = (List) entry.getValue();
-              entry.setValue(Collections.unmodifiableList(value));
-            }
-          }
-        }
-        super.makeImmutable();
-      }
-    };
-  }
-
-  /**
-   * Creates a new instance for testing.
-   *
-   * @param arraySize The size of the entry array containing the
-   *        lexicographically smallest mappings.
-   */
-  static <K extends Comparable<K>, V> SmallSortedMap<K, V> newInstanceForTest(
-      int arraySize) {
-    return new SmallSortedMap<K, V>(arraySize);
-  }
-
-  private final int maxArraySize;
-  // The "entry array" is actually a List because generic arrays are not
-  // allowed. ArrayList also nicely handles the entry shifting on inserts and
-  // removes.
-  private List<Entry> entryList;
-  private Map<K, V> overflowEntries;
-  private boolean isImmutable;
-  // The EntrySet is a stateless view of the Map. It's initialized the first
-  // time it is requested and reused henceforth.
-  private volatile EntrySet lazyEntrySet;
-
-  /**
-   * @code arraySize Size of the array in which the lexicographically smallest
-   *       mappings are stored. (i.e. the {@code k} referred to in the class
-   *       documentation).
-   */
-  private SmallSortedMap(int arraySize) {
-    this.maxArraySize = arraySize;
-    this.entryList = Collections.emptyList();
-    this.overflowEntries = Collections.emptyMap();
-  }
-
-  /** Make this map immutable from this point forward. */
-  public void makeImmutable() {
-    if (!isImmutable) {
-      // Note: There's no need to wrap the entryList in an unmodifiableList
-      // because none of the list's accessors are exposed. The iterator() of
-      // overflowEntries, on the other hand, is exposed so it must be made
-      // unmodifiable.
-      overflowEntries = overflowEntries.isEmpty() ?
-          Collections.<K, V>emptyMap() :
-          Collections.unmodifiableMap(overflowEntries);
-      isImmutable = true;
-    }
-  }
-
-  /** @return Whether {@link #makeImmutable()} has been called. */
-  public boolean isImmutable() {
-    return isImmutable;
-  }
-
-  /** @return The number of entries in the entry array. */
-  public int getNumArrayEntries() {
-    return entryList.size();
-  }
-
-  /** @return The array entry at the given {@code index}. */
-  public Map.Entry<K, V> getArrayEntryAt(int index) {
-    return entryList.get(index);
-  }
-
-  /** @return There number of overflow entries. */
-  public int getNumOverflowEntries() {
-    return overflowEntries.size();
-  }
-
-  /** @return An iterable over the overflow entries. */
-  public Iterable<Map.Entry<K, V>> getOverflowEntries() {
-    return overflowEntries.isEmpty() ?
-        EmptySet.<Map.Entry<K, V>>iterable() :
-        overflowEntries.entrySet();
-  }
-
-
-  @Override
-  public int size() {
-    return entryList.size() + overflowEntries.size();
-  }
-
-  /**
-   * The implementation throws a {@code ClassCastException} if o is not an
-   * object of type {@code K}.
-   *
-   * {@inheritDoc}
-   */
-  @Override
-  public boolean containsKey(Object o) {
-    @SuppressWarnings("unchecked")
-    final K key = (K) o;
-    return binarySearchInArray(key) >= 0 || overflowEntries.containsKey(key);
-  }
-
-  /**
-   * The implementation throws a {@code ClassCastException} if o is not an
-   * object of type {@code K}.
-   *
-   * {@inheritDoc}
-   */
-  @Override
-  public V get(Object o) {
-    @SuppressWarnings("unchecked")
-    final K key = (K) o;
-    final int index = binarySearchInArray(key);
-    if (index >= 0) {
-      return entryList.get(index).getValue();
-    }
-    return overflowEntries.get(key);
-  }
-
-  @Override
-  public V put(K key, V value) {
-    checkMutable();
-    final int index = binarySearchInArray(key);
-    if (index >= 0) {
-      // Replace existing array entry.
-      return entryList.get(index).setValue(value);
-    }
-    ensureEntryArrayMutable();
-    final int insertionPoint = -(index + 1);
-    if (insertionPoint >= maxArraySize) {
-      // Put directly in overflow.
-      return getOverflowEntriesMutable().put(key, value);
-    }
-    // Insert new Entry in array.
-    if (entryList.size() == maxArraySize) {
-      // Shift the last array entry into overflow.
-      final Entry lastEntryInArray = entryList.remove(maxArraySize - 1);
-      getOverflowEntriesMutable().put(lastEntryInArray.getKey(),
-                                      lastEntryInArray.getValue());
-    }
-    entryList.add(insertionPoint, new Entry(key, value));
-    return null;
-  }
-
-  @Override
-  public void clear() {
-    checkMutable();
-    if (!entryList.isEmpty()) {
-      entryList.clear();
-    }
-    if (!overflowEntries.isEmpty()) {
-      overflowEntries.clear();
-    }
-  }
-
-  /**
-   * The implementation throws a {@code ClassCastException} if o is not an
-   * object of type {@code K}.
-   *
-   * {@inheritDoc}
-   */
-  @Override
-  public V remove(Object o) {
-    checkMutable();
-    @SuppressWarnings("unchecked")
-    final K key = (K) o;
-    final int index = binarySearchInArray(key);
-    if (index >= 0) {
-      return removeArrayEntryAt(index);
-    }
-    // overflowEntries might be Collections.unmodifiableMap(), so only
-    // call remove() if it is non-empty.
-    if (overflowEntries.isEmpty()) {
-      return null;
-    } else {
-      return overflowEntries.remove(key);
-    }
-  }
-
-  private V removeArrayEntryAt(int index) {
-    checkMutable();
-    final V removed = entryList.remove(index).getValue();
-    if (!overflowEntries.isEmpty()) {
-      // Shift the first entry in the overflow to be the last entry in the
-      // array.
-      final Iterator<Map.Entry<K, V>> iterator =
-          getOverflowEntriesMutable().entrySet().iterator();
-      entryList.add(new Entry(iterator.next()));
-      iterator.remove();
-    }
-    return removed;
-  }
-
-  /**
-   * @param key The key to find in the entry array.
-   * @return The returned integer position follows the same semantics as the
-   *     value returned by {@link java.util.Arrays#binarySearch()}.
-   */
-  private int binarySearchInArray(K key) {
-    int left = 0;
-    int right = entryList.size() - 1;
-
-    // Optimization: For the common case in which entries are added in
-    // ascending tag order, check the largest element in the array before
-    // doing a full binary search.
-    if (right >= 0) {
-      int cmp = key.compareTo(entryList.get(right).getKey());
-      if (cmp > 0) {
-        return -(right + 2);  // Insert point is after "right".
-      } else if (cmp == 0) {
-        return right;
-      }
-    }
-
-    while (left <= right) {
-      int mid = (left + right) / 2;
-      int cmp = key.compareTo(entryList.get(mid).getKey());
-      if (cmp < 0) {
-        right = mid - 1;
-      } else if (cmp > 0) {
-        left = mid + 1;
-      } else {
-        return mid;
-      }
-    }
-    return -(left + 1);
-  }
-
-  /**
-   * Similar to the AbstractMap implementation of {@code keySet()} and
-   * {@code values()}, the entry set is created the first time this method is
-   * called, and returned in response to all subsequent calls.
-   *
-   * {@inheritDoc}
-   */
-  @Override
-  public Set<Map.Entry<K, V>> entrySet() {
-    if (lazyEntrySet == null) {
-      lazyEntrySet = new EntrySet();
-    }
-    return lazyEntrySet;
-  }
-
-
-  /**
-   * @throws UnsupportedOperationException if {@link #makeImmutable()} has
-   *         has been called.
-   */
-  private void checkMutable() {
-    if (isImmutable) {
-      throw new UnsupportedOperationException();
-    }
-  }
-
-  /**
-   * @return a {@link SortedMap} to which overflow entries mappings can be
-   *         added or removed.
-   * @throws UnsupportedOperationException if {@link #makeImmutable()} has been
-   *         called.
-   */
-  @SuppressWarnings("unchecked")
-  private SortedMap<K, V> getOverflowEntriesMutable() {
-    checkMutable();
-    if (overflowEntries.isEmpty() && !(overflowEntries instanceof TreeMap)) {
-      overflowEntries = new TreeMap<K, V>();
-    }
-    return (SortedMap<K, V>) overflowEntries;
-  }
-
-  /**
-   * Lazily creates the entry list. Any code that adds to the list must first
-   * call this method.
-   */
-  private void ensureEntryArrayMutable() {
-    checkMutable();
-    if (entryList.isEmpty() && !(entryList instanceof ArrayList)) {
-      entryList = new ArrayList<Entry>(maxArraySize);
-    }
-  }
-
-  /**
-   * Entry implementation that implements Comparable in order to support
-   * binary search within the entry array. Also checks mutability in
-   * {@link #setValue()}.
-   */
-  private class Entry implements Map.Entry<K, V>, Comparable<Entry> {
-
-    private final K key;
-    private V value;
-
-    Entry(Map.Entry<K, V> copy) {
-      this(copy.getKey(), copy.getValue());
-    }
-
-    Entry(K key, V value) {
-      this.key = key;
-      this.value = value;
-    }
-
-    @Override
-    public K getKey() {
-      return key;
-    }
-
-    @Override
-    public V getValue() {
-      return value;
-    }
-
-    @Override
-    public int compareTo(Entry other) {
-      return getKey().compareTo(other.getKey());
-    }
-
-    @Override
-    public V setValue(V newValue) {
-      checkMutable();
-      final V oldValue = this.value;
-      this.value = newValue;
-      return oldValue;
-    }
-
-    @Override
-    public boolean equals(Object o) {
-      if (o == this) {
-        return true;
-      }
-      if (!(o instanceof Map.Entry)) {
-        return false;
-      }
-      @SuppressWarnings("unchecked")
-      Map.Entry<?, ?> other = (Map.Entry<?, ?>) o;
-      return equals(key, other.getKey()) && equals(value, other.getValue());
-    }
-
-    @Override
-    public int hashCode() {
-      return (key == null ? 0 : key.hashCode()) ^
-          (value == null ? 0 : value.hashCode());
-    }
-
-    @Override
-    public String toString() {
-      return key + "=" + value;
-    }
-
-    /** equals() that handles null values. */
-    private boolean equals(Object o1, Object o2) {
-      return o1 == null ? o2 == null : o1.equals(o2);
-    }
-  }
-
-  /**
-   * Stateless view of the entries in the field map.
-   */
-  private class EntrySet extends AbstractSet<Map.Entry<K, V>> {
-
-    @Override
-    public Iterator<Map.Entry<K, V>> iterator() {
-      return new EntryIterator();
-    }
-
-    @Override
-    public int size() {
-      return SmallSortedMap.this.size();
-    }
-
-    /**
-     * Throws a {@link ClassCastException} if o is not of the expected type.
-     *
-     * {@inheritDoc}
-     */
-    @Override
-    public boolean contains(Object o) {
-      @SuppressWarnings("unchecked")
-      final Map.Entry<K, V> entry = (Map.Entry<K, V>) o;
-      final V existing = get(entry.getKey());
-      final V value = entry.getValue();
-      return existing == value ||
-          (existing != null && existing.equals(value));
-    }
-
-    @Override
-    public boolean add(Map.Entry<K, V> entry) {
-      if (!contains(entry)) {
-        put(entry.getKey(), entry.getValue());
-        return true;
-      }
-      return false;
-    }
-
-    /**
-     * Throws a {@link ClassCastException} if o is not of the expected type.
-     *
-     * {@inheritDoc}
-     */
-    @Override
-    public boolean remove(Object o) {
-      @SuppressWarnings("unchecked")
-      final Map.Entry<K, V> entry = (Map.Entry<K, V>) o;
-      if (contains(entry)) {
-        SmallSortedMap.this.remove(entry.getKey());
-        return true;
-      }
-      return false;
-    }
-
-    @Override
-    public void clear() {
-      SmallSortedMap.this.clear();
-    }
-  }
-
-
-  /**
-   * Iterator implementation that switches from the entry array to the overflow
-   * entries appropriately.
-   */
-  private class EntryIterator implements Iterator<Map.Entry<K, V>> {
-
-    private int pos = -1;
-    private boolean nextCalledBeforeRemove;
-    private Iterator<Map.Entry<K, V>> lazyOverflowIterator;
-
-    @Override
-    public boolean hasNext() {
-      return (pos + 1) < entryList.size() ||
-          getOverflowIterator().hasNext();
-    }
-
-    @Override
-    public Map.Entry<K, V> next() {
-      nextCalledBeforeRemove = true;
-      // Always increment pos so that we know whether the last returned value
-      // was from the array or from overflow.
-      if (++pos < entryList.size()) {
-        return entryList.get(pos);
-      }
-      return getOverflowIterator().next();
-    }
-
-    @Override
-    public void remove() {
-      if (!nextCalledBeforeRemove) {
-        throw new IllegalStateException("remove() was called before next()");
-      }
-      nextCalledBeforeRemove = false;
-      checkMutable();
-
-      if (pos < entryList.size()) {
-        removeArrayEntryAt(pos--);
-      } else {
-        getOverflowIterator().remove();
-      }
-    }
-
-    /**
-     * It is important to create the overflow iterator only after the array
-     * entries have been iterated over because the overflow entry set changes
-     * when the client calls remove() on the array entries, which invalidates
-     * any existing iterators.
-     */
-    private Iterator<Map.Entry<K, V>> getOverflowIterator() {
-      if (lazyOverflowIterator == null) {
-        lazyOverflowIterator = overflowEntries.entrySet().iterator();
-      }
-      return lazyOverflowIterator;
-    }
-  }
-
-  /**
-   * Helper class that holds immutable instances of an Iterable/Iterator that
-   * we return when the overflow entries is empty. This eliminates the creation
-   * of an Iterator object when there is nothing to iterate over.
-   */
-  private static class EmptySet {
-
-    private static final Iterator<Object> ITERATOR =
-        new Iterator<Object>() {
-          @Override
-          public boolean hasNext() {
-            return false;
-          }
-          @Override
-          public Object next() {
-            throw new NoSuchElementException();
-          }
-          @Override
-          public void remove() {
-            throw new UnsupportedOperationException();
-          }
-        };
-
-    private static final Iterable<Object> ITERABLE =
-        new Iterable<Object>() {
-          @Override
-          public Iterator<Object> iterator() {
-            return ITERATOR;
-          }
-        };
-
-    @SuppressWarnings("unchecked")
-    static <T> Iterable<T> iterable() {
-      return (Iterable<T>) ITERABLE;
-    }
-  }
-
-  @Override
-  public boolean equals(Object o) {
-    if (this == o) {
-      return true;
-    }
-
-    if (!(o instanceof SmallSortedMap)) {
-      return super.equals(o);
-    }
-
-    SmallSortedMap<?, ?> other = (SmallSortedMap<?, ?>) o;
-    final int size = size();
-    if (size != other.size()) {
-      return false;
-    }
-
-    // Best effort try to avoid allocating an entry set.
-    final int numArrayEntries = getNumArrayEntries();
-    if (numArrayEntries != other.getNumArrayEntries()) {
-      return entrySet().equals(other.entrySet());
-    }
-
-    for (int i = 0; i < numArrayEntries; i++) {
-      if (!getArrayEntryAt(i).equals(other.getArrayEntryAt(i))) {
-        return false;
-      }
-    }
-
-    if (numArrayEntries != size) {
-      return overflowEntries.equals(other.overflowEntries);
-    }
-
-
-    return true;
-  }
-
-  @Override
-  public int hashCode() {
-    int h = 0;
-    final int listSize = getNumArrayEntries();
-    for (int i = 0; i < listSize; i++) {
-      h += entryList.get(i).hashCode();
-    }
-    // Avoid the iterator allocation if possible.
-    if (getNumOverflowEntries() > 0) {
-      h += overflowEntries.hashCode();
-    }
-    return h;
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContext.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContext.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContext.java
deleted file mode 100644
index ab4938c..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContext.java
+++ /dev/null
@@ -1,539 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/source_context.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * `SourceContext` represents information about the source of a
- * protobuf element, like the file in which it is defined.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.SourceContext}
- */
-public  final class SourceContext extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.SourceContext)
-    SourceContextOrBuilder {
-  // Use SourceContext.newBuilder() to construct.
-  private SourceContext(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private SourceContext() {
-    fileName_ = "";
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private SourceContext(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            fileName_ = s;
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.internal_static_google_protobuf_SourceContext_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.internal_static_google_protobuf_SourceContext_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder.class);
-  }
-
-  public static final int FILE_NAME_FIELD_NUMBER = 1;
-  private volatile java.lang.Object fileName_;
-  /**
-   * <pre>
-   * The path-qualified name of the .proto file that contained the associated
-   * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-   * </pre>
-   *
-   * <code>string file_name = 1;</code>
-   */
-  public java.lang.String getFileName() {
-    java.lang.Object ref = fileName_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      fileName_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * The path-qualified name of the .proto file that contained the associated
-   * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-   * </pre>
-   *
-   * <code>string file_name = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getFileNameBytes() {
-    java.lang.Object ref = fileName_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      fileName_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getFileNameBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, fileName_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getFileNameBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, fileName_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext) obj;
-
-    boolean result = true;
-    result = result && getFileName()
-        .equals(other.getFileName());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + FILE_NAME_FIELD_NUMBER;
-    hash = (53 * hash) + getFileName().hashCode();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * `SourceContext` represents information about the source of a
-   * protobuf element, like the file in which it is defined.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.SourceContext}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.SourceContext)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.internal_static_google_protobuf_SourceContext_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.internal_static_google_protobuf_SourceContext_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      fileName_ = "";
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.internal_static_google_protobuf_SourceContext_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext(this);
-      result.fileName_ = fileName_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext.getDefaultInstance()) return this;
-      if (!other.getFileName().isEmpty()) {
-        fileName_ = other.fileName_;
-        onChanged();
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private java.lang.Object fileName_ = "";
-    /**
-     * <pre>
-     * The path-qualified name of the .proto file that contained the associated
-     * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-     * </pre>
-     *
-     * <code>string file_name = 1;</code>
-     */
-    public java.lang.String getFileName() {
-      java.lang.Object ref = fileName_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        fileName_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The path-qualified name of the .proto file that contained the associated
-     * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-     * </pre>
-     *
-     * <code>string file_name = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getFileNameBytes() {
-      java.lang.Object ref = fileName_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        fileName_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * The path-qualified name of the .proto file that contained the associated
-     * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-     * </pre>
-     *
-     * <code>string file_name = 1;</code>
-     */
-    public Builder setFileName(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      fileName_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The path-qualified name of the .proto file that contained the associated
-     * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-     * </pre>
-     *
-     * <code>string file_name = 1;</code>
-     */
-    public Builder clearFileName() {
-      
-      fileName_ = getDefaultInstance().getFileName();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The path-qualified name of the .proto file that contained the associated
-     * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-     * </pre>
-     *
-     * <code>string file_name = 1;</code>
-     */
-    public Builder setFileNameBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      fileName_ = value;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.SourceContext)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.SourceContext)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SourceContext>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<SourceContext>() {
-    public SourceContext parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new SourceContext(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SourceContext> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<SourceContext> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContextOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContextOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContextOrBuilder.java
deleted file mode 100644
index c35eceb..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContextOrBuilder.java
+++ /dev/null
@@ -1,29 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/source_context.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface SourceContextOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.SourceContext)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The path-qualified name of the .proto file that contained the associated
-   * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-   * </pre>
-   *
-   * <code>string file_name = 1;</code>
-   */
-  java.lang.String getFileName();
-  /**
-   * <pre>
-   * The path-qualified name of the .proto file that contained the associated
-   * protobuf element.  For example: `"google/protobuf/source_context.proto"`.
-   * </pre>
-   *
-   * <code>string file_name = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getFileNameBytes();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContextProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContextProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContextProto.java
deleted file mode 100644
index 48f2e3f..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/SourceContextProto.java
+++ /dev/null
@@ -1,60 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/source_context.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class SourceContextProto {
-  private SourceContextProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_SourceContext_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_SourceContext_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n$google/protobuf/source_context.proto\022\017" +
-      "google.protobuf\"\"\n\rSourceContext\022\021\n\tfile" +
-      "_name\030\001 \001(\tB\225\001\n\023com.google.protobufB\022Sou" +
-      "rceContextProtoP\001ZAgoogle.golang.org/gen" +
-      "proto/protobuf/source_context;source_con" +
-      "text\242\002\003GPB\252\002\036Google.Protobuf.WellKnownTy" +
-      "pesb\006proto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_google_protobuf_SourceContext_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_SourceContext_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_SourceContext_descriptor,
-        new java.lang.String[] { "FileName", });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}


[45/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java
----------------------------------------------------------------------
diff --git a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java b/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java
deleted file mode 100644
index 7ba5b8e..0000000
--- a/hbase-endpoint/src/main/java/org/apache/hadoop/hbase/coprocessor/protobuf/generated/IncrementCounterProcessorTestProtos.java
+++ /dev/null
@@ -1,4059 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: IncrementCounterProcessor.proto
-
-package org.apache.hadoop.hbase.coprocessor.protobuf.generated;
-
-public final class IncrementCounterProcessorTestProtos {
-  private IncrementCounterProcessorTestProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface IncCounterProcessorRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required bytes row = 1;
-    /**
-     * <code>required bytes row = 1;</code>
-     */
-    boolean hasRow();
-    /**
-     * <code>required bytes row = 1;</code>
-     */
-    com.google.protobuf.ByteString getRow();
-
-    // required int32 counter = 2;
-    /**
-     * <code>required int32 counter = 2;</code>
-     */
-    boolean hasCounter();
-    /**
-     * <code>required int32 counter = 2;</code>
-     */
-    int getCounter();
-  }
-  /**
-   * Protobuf type {@code IncCounterProcessorRequest}
-   */
-  public static final class IncCounterProcessorRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements IncCounterProcessorRequestOrBuilder {
-    // Use IncCounterProcessorRequest.newBuilder() to construct.
-    private IncCounterProcessorRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private IncCounterProcessorRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final IncCounterProcessorRequest defaultInstance;
-    public static IncCounterProcessorRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public IncCounterProcessorRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private IncCounterProcessorRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              row_ = input.readBytes();
-              break;
-            }
-            case 16: {
-              bitField0_ |= 0x00000002;
-              counter_ = input.readInt32();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<IncCounterProcessorRequest> PARSER =
-        new com.google.protobuf.AbstractParser<IncCounterProcessorRequest>() {
-      public IncCounterProcessorRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new IncCounterProcessorRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<IncCounterProcessorRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required bytes row = 1;
-    public static final int ROW_FIELD_NUMBER = 1;
-    private com.google.protobuf.ByteString row_;
-    /**
-     * <code>required bytes row = 1;</code>
-     */
-    public boolean hasRow() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required bytes row = 1;</code>
-     */
-    public com.google.protobuf.ByteString getRow() {
-      return row_;
-    }
-
-    // required int32 counter = 2;
-    public static final int COUNTER_FIELD_NUMBER = 2;
-    private int counter_;
-    /**
-     * <code>required int32 counter = 2;</code>
-     */
-    public boolean hasCounter() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>required int32 counter = 2;</code>
-     */
-    public int getCounter() {
-      return counter_;
-    }
-
-    private void initFields() {
-      row_ = com.google.protobuf.ByteString.EMPTY;
-      counter_ = 0;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasRow()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!hasCounter()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, row_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeInt32(2, counter_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, row_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt32Size(2, counter_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest) obj;
-
-      boolean result = true;
-      result = result && (hasRow() == other.hasRow());
-      if (hasRow()) {
-        result = result && getRow()
-            .equals(other.getRow());
-      }
-      result = result && (hasCounter() == other.hasCounter());
-      if (hasCounter()) {
-        result = result && (getCounter()
-            == other.getCounter());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasRow()) {
-        hash = (37 * hash) + ROW_FIELD_NUMBER;
-        hash = (53 * hash) + getRow().hashCode();
-      }
-      if (hasCounter()) {
-        hash = (37 * hash) + COUNTER_FIELD_NUMBER;
-        hash = (53 * hash) + getCounter();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code IncCounterProcessorRequest}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        row_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        counter_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.row_ = row_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.counter_ = counter_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest.getDefaultInstance()) return this;
-        if (other.hasRow()) {
-          setRow(other.getRow());
-        }
-        if (other.hasCounter()) {
-          setCounter(other.getCounter());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasRow()) {
-          
-          return false;
-        }
-        if (!hasCounter()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required bytes row = 1;
-      private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>required bytes row = 1;</code>
-       */
-      public boolean hasRow() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required bytes row = 1;</code>
-       */
-      public com.google.protobuf.ByteString getRow() {
-        return row_;
-      }
-      /**
-       * <code>required bytes row = 1;</code>
-       */
-      public Builder setRow(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        row_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required bytes row = 1;</code>
-       */
-      public Builder clearRow() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        row_ = getDefaultInstance().getRow();
-        onChanged();
-        return this;
-      }
-
-      // required int32 counter = 2;
-      private int counter_ ;
-      /**
-       * <code>required int32 counter = 2;</code>
-       */
-      public boolean hasCounter() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>required int32 counter = 2;</code>
-       */
-      public int getCounter() {
-        return counter_;
-      }
-      /**
-       * <code>required int32 counter = 2;</code>
-       */
-      public Builder setCounter(int value) {
-        bitField0_ |= 0x00000002;
-        counter_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required int32 counter = 2;</code>
-       */
-      public Builder clearCounter() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        counter_ = 0;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:IncCounterProcessorRequest)
-    }
-
-    static {
-      defaultInstance = new IncCounterProcessorRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:IncCounterProcessorRequest)
-  }
-
-  public interface IncCounterProcessorResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required int32 response = 1;
-    /**
-     * <code>required int32 response = 1;</code>
-     */
-    boolean hasResponse();
-    /**
-     * <code>required int32 response = 1;</code>
-     */
-    int getResponse();
-  }
-  /**
-   * Protobuf type {@code IncCounterProcessorResponse}
-   */
-  public static final class IncCounterProcessorResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements IncCounterProcessorResponseOrBuilder {
-    // Use IncCounterProcessorResponse.newBuilder() to construct.
-    private IncCounterProcessorResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private IncCounterProcessorResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final IncCounterProcessorResponse defaultInstance;
-    public static IncCounterProcessorResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public IncCounterProcessorResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private IncCounterProcessorResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              response_ = input.readInt32();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<IncCounterProcessorResponse> PARSER =
-        new com.google.protobuf.AbstractParser<IncCounterProcessorResponse>() {
-      public IncCounterProcessorResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new IncCounterProcessorResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<IncCounterProcessorResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required int32 response = 1;
-    public static final int RESPONSE_FIELD_NUMBER = 1;
-    private int response_;
-    /**
-     * <code>required int32 response = 1;</code>
-     */
-    public boolean hasResponse() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required int32 response = 1;</code>
-     */
-    public int getResponse() {
-      return response_;
-    }
-
-    private void initFields() {
-      response_ = 0;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasResponse()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeInt32(1, response_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt32Size(1, response_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse) obj;
-
-      boolean result = true;
-      result = result && (hasResponse() == other.hasResponse());
-      if (hasResponse()) {
-        result = result && (getResponse()
-            == other.getResponse());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasResponse()) {
-        hash = (37 * hash) + RESPONSE_FIELD_NUMBER;
-        hash = (53 * hash) + getResponse();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code IncCounterProcessorResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        response_ = 0;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_IncCounterProcessorResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.response_ = response_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse.getDefaultInstance()) return this;
-        if (other.hasResponse()) {
-          setResponse(other.getResponse());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasResponse()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.IncCounterProcessorResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required int32 response = 1;
-      private int response_ ;
-      /**
-       * <code>required int32 response = 1;</code>
-       */
-      public boolean hasResponse() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required int32 response = 1;</code>
-       */
-      public int getResponse() {
-        return response_;
-      }
-      /**
-       * <code>required int32 response = 1;</code>
-       */
-      public Builder setResponse(int value) {
-        bitField0_ |= 0x00000001;
-        response_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required int32 response = 1;</code>
-       */
-      public Builder clearResponse() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        response_ = 0;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:IncCounterProcessorResponse)
-    }
-
-    static {
-      defaultInstance = new IncCounterProcessorResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:IncCounterProcessorResponse)
-  }
-
-  public interface FriendsOfFriendsProcessorRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required bytes person = 1;
-    /**
-     * <code>required bytes person = 1;</code>
-     */
-    boolean hasPerson();
-    /**
-     * <code>required bytes person = 1;</code>
-     */
-    com.google.protobuf.ByteString getPerson();
-
-    // required bytes row = 2;
-    /**
-     * <code>required bytes row = 2;</code>
-     */
-    boolean hasRow();
-    /**
-     * <code>required bytes row = 2;</code>
-     */
-    com.google.protobuf.ByteString getRow();
-
-    // repeated string result = 3;
-    /**
-     * <code>repeated string result = 3;</code>
-     */
-    java.util.List<java.lang.String>
-    getResultList();
-    /**
-     * <code>repeated string result = 3;</code>
-     */
-    int getResultCount();
-    /**
-     * <code>repeated string result = 3;</code>
-     */
-    java.lang.String getResult(int index);
-    /**
-     * <code>repeated string result = 3;</code>
-     */
-    com.google.protobuf.ByteString
-        getResultBytes(int index);
-  }
-  /**
-   * Protobuf type {@code FriendsOfFriendsProcessorRequest}
-   */
-  public static final class FriendsOfFriendsProcessorRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements FriendsOfFriendsProcessorRequestOrBuilder {
-    // Use FriendsOfFriendsProcessorRequest.newBuilder() to construct.
-    private FriendsOfFriendsProcessorRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private FriendsOfFriendsProcessorRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final FriendsOfFriendsProcessorRequest defaultInstance;
-    public static FriendsOfFriendsProcessorRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public FriendsOfFriendsProcessorRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private FriendsOfFriendsProcessorRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              person_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              bitField0_ |= 0x00000002;
-              row_ = input.readBytes();
-              break;
-            }
-            case 26: {
-              if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-                result_ = new com.google.protobuf.LazyStringArrayList();
-                mutable_bitField0_ |= 0x00000004;
-              }
-              result_.add(input.readBytes());
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
-          result_ = new com.google.protobuf.UnmodifiableLazyStringList(result_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<FriendsOfFriendsProcessorRequest> PARSER =
-        new com.google.protobuf.AbstractParser<FriendsOfFriendsProcessorRequest>() {
-      public FriendsOfFriendsProcessorRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new FriendsOfFriendsProcessorRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<FriendsOfFriendsProcessorRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required bytes person = 1;
-    public static final int PERSON_FIELD_NUMBER = 1;
-    private com.google.protobuf.ByteString person_;
-    /**
-     * <code>required bytes person = 1;</code>
-     */
-    public boolean hasPerson() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required bytes person = 1;</code>
-     */
-    public com.google.protobuf.ByteString getPerson() {
-      return person_;
-    }
-
-    // required bytes row = 2;
-    public static final int ROW_FIELD_NUMBER = 2;
-    private com.google.protobuf.ByteString row_;
-    /**
-     * <code>required bytes row = 2;</code>
-     */
-    public boolean hasRow() {
-      return ((bitField0_ & 0x00000002) == 0x00000002);
-    }
-    /**
-     * <code>required bytes row = 2;</code>
-     */
-    public com.google.protobuf.ByteString getRow() {
-      return row_;
-    }
-
-    // repeated string result = 3;
-    public static final int RESULT_FIELD_NUMBER = 3;
-    private com.google.protobuf.LazyStringList result_;
-    /**
-     * <code>repeated string result = 3;</code>
-     */
-    public java.util.List<java.lang.String>
-        getResultList() {
-      return result_;
-    }
-    /**
-     * <code>repeated string result = 3;</code>
-     */
-    public int getResultCount() {
-      return result_.size();
-    }
-    /**
-     * <code>repeated string result = 3;</code>
-     */
-    public java.lang.String getResult(int index) {
-      return result_.get(index);
-    }
-    /**
-     * <code>repeated string result = 3;</code>
-     */
-    public com.google.protobuf.ByteString
-        getResultBytes(int index) {
-      return result_.getByteString(index);
-    }
-
-    private void initFields() {
-      person_ = com.google.protobuf.ByteString.EMPTY;
-      row_ = com.google.protobuf.ByteString.EMPTY;
-      result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasPerson()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      if (!hasRow()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, person_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeBytes(2, row_);
-      }
-      for (int i = 0; i < result_.size(); i++) {
-        output.writeBytes(3, result_.getByteString(i));
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, person_);
-      }
-      if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(2, row_);
-      }
-      {
-        int dataSize = 0;
-        for (int i = 0; i < result_.size(); i++) {
-          dataSize += com.google.protobuf.CodedOutputStream
-            .computeBytesSizeNoTag(result_.getByteString(i));
-        }
-        size += dataSize;
-        size += 1 * getResultList().size();
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest) obj;
-
-      boolean result = true;
-      result = result && (hasPerson() == other.hasPerson());
-      if (hasPerson()) {
-        result = result && getPerson()
-            .equals(other.getPerson());
-      }
-      result = result && (hasRow() == other.hasRow());
-      if (hasRow()) {
-        result = result && getRow()
-            .equals(other.getRow());
-      }
-      result = result && getResultList()
-          .equals(other.getResultList());
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasPerson()) {
-        hash = (37 * hash) + PERSON_FIELD_NUMBER;
-        hash = (53 * hash) + getPerson().hashCode();
-      }
-      if (hasRow()) {
-        hash = (37 * hash) + ROW_FIELD_NUMBER;
-        hash = (53 * hash) + getRow().hashCode();
-      }
-      if (getResultCount() > 0) {
-        hash = (37 * hash) + RESULT_FIELD_NUMBER;
-        hash = (53 * hash) + getResultList().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code FriendsOfFriendsProcessorRequest}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        person_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        row_ = com.google.protobuf.ByteString.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000002);
-        result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000004);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.person_ = person_;
-        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
-          to_bitField0_ |= 0x00000002;
-        }
-        result.row_ = row_;
-        if (((bitField0_ & 0x00000004) == 0x00000004)) {
-          result_ = new com.google.protobuf.UnmodifiableLazyStringList(
-              result_);
-          bitField0_ = (bitField0_ & ~0x00000004);
-        }
-        result.result_ = result_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest.getDefaultInstance()) return this;
-        if (other.hasPerson()) {
-          setPerson(other.getPerson());
-        }
-        if (other.hasRow()) {
-          setRow(other.getRow());
-        }
-        if (!other.result_.isEmpty()) {
-          if (result_.isEmpty()) {
-            result_ = other.result_;
-            bitField0_ = (bitField0_ & ~0x00000004);
-          } else {
-            ensureResultIsMutable();
-            result_.addAll(other.result_);
-          }
-          onChanged();
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasPerson()) {
-          
-          return false;
-        }
-        if (!hasRow()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required bytes person = 1;
-      private com.google.protobuf.ByteString person_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>required bytes person = 1;</code>
-       */
-      public boolean hasPerson() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required bytes person = 1;</code>
-       */
-      public com.google.protobuf.ByteString getPerson() {
-        return person_;
-      }
-      /**
-       * <code>required bytes person = 1;</code>
-       */
-      public Builder setPerson(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000001;
-        person_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required bytes person = 1;</code>
-       */
-      public Builder clearPerson() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        person_ = getDefaultInstance().getPerson();
-        onChanged();
-        return this;
-      }
-
-      // required bytes row = 2;
-      private com.google.protobuf.ByteString row_ = com.google.protobuf.ByteString.EMPTY;
-      /**
-       * <code>required bytes row = 2;</code>
-       */
-      public boolean hasRow() {
-        return ((bitField0_ & 0x00000002) == 0x00000002);
-      }
-      /**
-       * <code>required bytes row = 2;</code>
-       */
-      public com.google.protobuf.ByteString getRow() {
-        return row_;
-      }
-      /**
-       * <code>required bytes row = 2;</code>
-       */
-      public Builder setRow(com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  bitField0_ |= 0x00000002;
-        row_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required bytes row = 2;</code>
-       */
-      public Builder clearRow() {
-        bitField0_ = (bitField0_ & ~0x00000002);
-        row_ = getDefaultInstance().getRow();
-        onChanged();
-        return this;
-      }
-
-      // repeated string result = 3;
-      private com.google.protobuf.LazyStringList result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-      private void ensureResultIsMutable() {
-        if (!((bitField0_ & 0x00000004) == 0x00000004)) {
-          result_ = new com.google.protobuf.LazyStringArrayList(result_);
-          bitField0_ |= 0x00000004;
-         }
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public java.util.List<java.lang.String>
-          getResultList() {
-        return java.util.Collections.unmodifiableList(result_);
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public int getResultCount() {
-        return result_.size();
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public java.lang.String getResult(int index) {
-        return result_.get(index);
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public com.google.protobuf.ByteString
-          getResultBytes(int index) {
-        return result_.getByteString(index);
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public Builder setResult(
-          int index, java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureResultIsMutable();
-        result_.set(index, value);
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public Builder addResult(
-          java.lang.String value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureResultIsMutable();
-        result_.add(value);
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public Builder addAllResult(
-          java.lang.Iterable<java.lang.String> values) {
-        ensureResultIsMutable();
-        super.addAll(values, result_);
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public Builder clearResult() {
-        result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000004);
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>repeated string result = 3;</code>
-       */
-      public Builder addResultBytes(
-          com.google.protobuf.ByteString value) {
-        if (value == null) {
-    throw new NullPointerException();
-  }
-  ensureResultIsMutable();
-        result_.add(value);
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:FriendsOfFriendsProcessorRequest)
-    }
-
-    static {
-      defaultInstance = new FriendsOfFriendsProcessorRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:FriendsOfFriendsProcessorRequest)
-  }
-
-  public interface FriendsOfFriendsProcessorResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // repeated string result = 1;
-    /**
-     * <code>repeated string result = 1;</code>
-     */
-    java.util.List<java.lang.String>
-    getResultList();
-    /**
-     * <code>repeated string result = 1;</code>
-     */
-    int getResultCount();
-    /**
-     * <code>repeated string result = 1;</code>
-     */
-    java.lang.String getResult(int index);
-    /**
-     * <code>repeated string result = 1;</code>
-     */
-    com.google.protobuf.ByteString
-        getResultBytes(int index);
-  }
-  /**
-   * Protobuf type {@code FriendsOfFriendsProcessorResponse}
-   */
-  public static final class FriendsOfFriendsProcessorResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements FriendsOfFriendsProcessorResponseOrBuilder {
-    // Use FriendsOfFriendsProcessorResponse.newBuilder() to construct.
-    private FriendsOfFriendsProcessorResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private FriendsOfFriendsProcessorResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final FriendsOfFriendsProcessorResponse defaultInstance;
-    public static FriendsOfFriendsProcessorResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public FriendsOfFriendsProcessorResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private FriendsOfFriendsProcessorResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 10: {
-              if (!((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-                result_ = new com.google.protobuf.LazyStringArrayList();
-                mutable_bitField0_ |= 0x00000001;
-              }
-              result_.add(input.readBytes());
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        if (((mutable_bitField0_ & 0x00000001) == 0x00000001)) {
-          result_ = new com.google.protobuf.UnmodifiableLazyStringList(result_);
-        }
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<FriendsOfFriendsProcessorResponse> PARSER =
-        new com.google.protobuf.AbstractParser<FriendsOfFriendsProcessorResponse>() {
-      public FriendsOfFriendsProcessorResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new FriendsOfFriendsProcessorResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<FriendsOfFriendsProcessorResponse> getParserForType() {
-      return PARSER;
-    }
-
-    // repeated string result = 1;
-    public static final int RESULT_FIELD_NUMBER = 1;
-    private com.google.protobuf.LazyStringList result_;
-    /**
-     * <code>repeated string result = 1;</code>
-     */
-    public java.util.List<java.lang.String>
-        getResultList() {
-      return result_;
-    }
-    /**
-     * <code>repeated string result = 1;</code>
-     */
-    public int getResultCount() {
-      return result_.size();
-    }
-    /**
-     * <code>repeated string result = 1;</code>
-     */
-    public java.lang.String getResult(int index) {
-      return result_.get(index);
-    }
-    /**
-     * <code>repeated string result = 1;</code>
-     */
-    public com.google.protobuf.ByteString
-        getResultBytes(int index) {
-      return result_.getByteString(index);
-    }
-
-    private void initFields() {
-      result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      for (int i = 0; i < result_.size(); i++) {
-        output.writeBytes(1, result_.getByteString(i));
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      {
-        int dataSize = 0;
-        for (int i = 0; i < result_.size(); i++) {
-          dataSize += com.google.protobuf.CodedOutputStream
-            .computeBytesSizeNoTag(result_.getByteString(i));
-        }
-        size += dataSize;
-        size += 1 * getResultList().size();
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse other = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse) obj;
-
-      boolean result = true;
-      result = result && getResultList()
-          .equals(other.getResultList());
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (getResultCount() > 0) {
-        hash = (37 * hash) + RESULT_FIELD_NUMBER;
-        hash = (53 * hash) + getResultList().hashCode();
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code FriendsOfFriendsProcessorResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.class, org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.internal_static_FriendsOfFriendsProcessorResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse build() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse result = new org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse(this);
-        int from_bitField0_ = bitField0_;
-        if (((bitField0_ & 0x00000001) == 0x00000001)) {
-          result_ = new com.google.protobuf.UnmodifiableLazyStringList(
-              result_);
-          bitField0_ = (bitField0_ & ~0x00000001);
-        }
-        result.result_ = result_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse.getDefaultInstance()) return this;
-        if (!other.result_.isEmpty()) {
-          if (result_.isEmpty()) {
-            result_ = other.result_;
-            bitField0_ = (bitField0_ & ~0x00000001);
-          } else {
-            ensureResultIsMutable();
-            result_.addAll(other.result_);
-          }
-          onChanged();
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.FriendsOfFriendsProcessorResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // repeated string result = 1;
-      private com.google.protobuf.LazyStringList result_ = com.google.protobuf.LazyStringArrayList.EMPTY;
-      private void ensureResultIsMutable() {
-        if (!((bitField0_ & 0x00000001) == 0x00000001)) {
-          result_ = new com.google.protobuf.LazyStringArrayList(result_);
-       

<TRUNCATED>

[42/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-examples/README.txt
----------------------------------------------------------------------
diff --git a/hbase-examples/README.txt b/hbase-examples/README.txt
index c47ed4f..22d1103 100644
--- a/hbase-examples/README.txt
+++ b/hbase-examples/README.txt
@@ -63,27 +63,8 @@ Example code.
       3. Execute {./DemoClient}.
 
 ON PROTOBUFS
-This maven module has  protobuf definition files ('.protos') used by hbase
-Coprocessor Endpoints examples including tests. Coprocessor
-Endpoints are meant to be standalone, independent code not reliant on hbase
-internals. They define their Service using protobuf. The protobuf version
-they use can be distinct from that used by HBase internally since HBase started
-shading its protobuf references. Endpoints have no access to the shaded protobuf
-hbase uses. They do have access to the content of hbase-protocol -- the
-.protos found in here -- but avoid using as much of this as you can as it is
-liable to change.
+This maven module has core protobuf definition files ('.protos') used by hbase
+examples. 
 
-Generation of java files from protobuf .proto files included here is done apart
-from the build. Run the generation whenever you make changes to the .orotos files
-and then check in the produced java (The reasoning is that change is infrequent
-so why pay the price of generating files anew on each build.
-
-To generate java files from protos run:
-
- $ mvn compile -Dcompile-protobuf
-or
- $ mvn compile -Pcompile-protobuf
-
-After you've done the above, check it and then check in changes (or post a patch
-on a JIRA with your definition file changes and the generated files). Be careful
-to notice new files and files removed and do appropriate git rm/adds.
+Generation of java files from protobuf .proto files included here is done as
+part of the build.

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-examples/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-examples/pom.xml b/hbase-examples/pom.xml
index 0eee552..9746b9a 100644
--- a/hbase-examples/pom.xml
+++ b/hbase-examples/pom.xml
@@ -29,7 +29,7 @@
   <artifactId>hbase-examples</artifactId>
   <name>Apache HBase - Examples</name>
   <description>Examples of HBase usage</description>
-<!--REMOVE-->
+  <!--REMOVE-->
   <build>
     <plugins>
       <plugin>
@@ -46,20 +46,33 @@
           <skipAssembly>true</skipAssembly>
         </configuration>
       </plugin>
-        <plugin>
-            <artifactId>maven-surefire-plugin</artifactId>
-            <version>${surefire.version}</version>
-            <configuration>
-                <!-- Have to set the groups here because we only do
+      <plugin>
+        <artifactId>maven-surefire-plugin</artifactId>
+        <version>${surefire.version}</version>
+        <configuration>
+          <!-- Have to set the groups here because we only do
                     split tests in this package, so groups on live in this module -->
-                <groups>${surefire.firstPartGroups}</groups>
-            </configuration>
-        </plugin>
-        <!-- Make a jar and put the sources in the jar -->
-        <plugin>
-          <groupId>org.apache.maven.plugins</groupId>
-          <artifactId>maven-source-plugin</artifactId>
-        </plugin>
+          <groups>${surefire.firstPartGroups}</groups>
+        </configuration>
+      </plugin>
+      <!-- Make a jar and put the sources in the jar -->
+      <plugin>
+        <groupId>org.apache.maven.plugins</groupId>
+        <artifactId>maven-source-plugin</artifactId>
+      </plugin>
+      <plugin>
+        <groupId>org.xolstice.maven.plugins</groupId>
+        <artifactId>protobuf-maven-plugin</artifactId>
+        <executions>
+          <execution>
+            <id>compile-protoc</id>
+            <phase>generate-sources</phase>
+            <goals>
+              <goal>compile</goal>
+            </goals>
+          </execution>
+        </executions>
+      </plugin>
     </plugins>
     <pluginManagement>
       <plugins>
@@ -81,7 +94,7 @@
                     </goals>
                   </pluginExecutionFilter>
                   <action>
-                    <ignore></ignore>
+                    <ignore/>
                   </action>
                 </pluginExecution>
                 <pluginExecution>
@@ -94,7 +107,7 @@
                     </goals>
                   </pluginExecutionFilter>
                   <action>
-                    <ignore></ignore>
+                    <ignore/>
                   </action>
                 </pluginExecution>
               </pluginExecutions>
@@ -116,16 +129,16 @@
       <scope>test</scope>
     </dependency>
     <dependency>
-        <groupId>org.apache.hbase</groupId>
-        <artifactId>hbase-common</artifactId>
+      <groupId>org.apache.hbase</groupId>
+      <artifactId>hbase-common</artifactId>
     </dependency>
     <dependency>
-        <groupId>org.apache.hbase</groupId>
-        <artifactId>hbase-protocol</artifactId>
+      <groupId>org.apache.hbase</groupId>
+      <artifactId>hbase-protocol</artifactId>
     </dependency>
     <dependency>
-        <groupId>org.apache.hbase</groupId>
-        <artifactId>hbase-client</artifactId>
+      <groupId>org.apache.hbase</groupId>
+      <artifactId>hbase-client</artifactId>
     </dependency>
     <dependency>
       <groupId>org.apache.hbase</groupId>
@@ -140,11 +153,10 @@
       <artifactId>hbase-thrift</artifactId>
     </dependency>
     <dependency>
-        <groupId>org.apache.hbase</groupId>
-        <artifactId>hbase-testing-util</artifactId>
-        <scope>test</scope>
+      <groupId>org.apache.hbase</groupId>
+      <artifactId>hbase-testing-util</artifactId>
+      <scope>test</scope>
     </dependency>
-
     <dependency>
       <groupId>org.apache.thrift</groupId>
       <artifactId>libthrift</artifactId>
@@ -161,158 +173,126 @@
       <groupId>com.google.protobuf</groupId>
       <artifactId>protobuf-java</artifactId>
     </dependency>
- </dependencies>
- <profiles>
-     <!-- Skip the tests in this module -->
-     <profile>
-         <id>skipExamplesTests</id>
-         <activation>
-             <property>
-                 <name>skipExamplesTests</name>
-             </property>
-         </activation>
-         <properties>
-             <surefire.skipFirstPart>true</surefire.skipFirstPart>
-             <surefire.skipSecondPart>true</surefire.skipSecondPart>
-         </properties>
-     </profile>
-        <profile>
-          <id>compile-protobuf</id>
-          <activation>
-            <property>
-              <name>compile-protobuf</name>
-            </property>
-          </activation>
-          <build>
-            <plugins>
-              <plugin>
-                <groupId>org.xolstice.maven.plugins</groupId>
-                <artifactId>protobuf-maven-plugin</artifactId>
-                <executions>
-                  <execution>
-                    <id>compile-protoc</id>
-                    <phase>generate-sources</phase>
-                    <goals>
-                      <goal>compile</goal>
-                    </goals>
-                  </execution>
-                </executions>
-              </plugin>
-            </plugins>
-          </build>
-        </profile>
-
-     <!-- Profiles for building against different hadoop versions -->
-     <!-- There are a lot of common dependencies used here, should investigate
-if we can combine these profiles somehow -->
-
-     <!-- profile for building against Hadoop 2.x.  This is the default -->
-     <profile>
-         <id>hadoop-2.0</id>
-         <activation>
-             <property>
-            <!--Below formatting for dev-support/generate-hadoopX-poms.sh-->
-            <!--h2--><name>!hadoop.profile</name>
-             </property>
-         </activation>
-         <dependencies>
-             <dependency>
-                 <groupId>org.apache.hadoop</groupId>
-                 <artifactId>hadoop-mapreduce-client-core</artifactId>
-                <exclusions>
-                  <exclusion>
-                    <groupId>com.google.guava</groupId>
-                    <artifactId>guava</artifactId>
-                  </exclusion>
-                </exclusions>
-             </dependency>
-             <dependency>
-                 <groupId>org.apache.hadoop</groupId>
-                 <artifactId>hadoop-common</artifactId>
-             </dependency>
-         </dependencies>
-         <build>
-             <plugins>
-                 <plugin>
-                     <artifactId>maven-dependency-plugin</artifactId>
-                     <executions>
-                         <execution>
-                             <id>create-mrapp-generated-classpath</id>
-                             <phase>generate-test-resources</phase>
-                             <goals>
-                                 <goal>build-classpath</goal>
-                             </goals>
-                             <configuration>
-                                 <!-- needed to run the unit test for DS to generate
+  </dependencies>
+  <profiles>
+    <!-- Skip the tests in this module -->
+    <profile>
+      <id>skipExamplesTests</id>
+      <activation>
+        <property>
+          <name>skipExamplesTests</name>
+        </property>
+      </activation>
+      <properties>
+        <surefire.skipFirstPart>true</surefire.skipFirstPart>
+        <surefire.skipSecondPart>true</surefire.skipSecondPart>
+      </properties>
+    </profile>
+    <!-- Profiles for building against different hadoop versions -->
+    <!-- There are a lot of common dependencies used here, should investigate
+         if we can combine these profiles somehow -->
+    <!-- profile for building against Hadoop 2.x.  This is the default -->
+    <profile>
+      <id>hadoop-2.0</id>
+      <activation>
+        <property>
+          <!--Below formatting for dev-support/generate-hadoopX-poms.sh-->
+          <!--h2-->
+          <name>!hadoop.profile</name>
+        </property>
+      </activation>
+      <dependencies>
+        <dependency>
+          <groupId>org.apache.hadoop</groupId>
+          <artifactId>hadoop-mapreduce-client-core</artifactId>
+        </dependency>
+        <dependency>
+          <groupId>org.apache.hadoop</groupId>
+          <artifactId>hadoop-common</artifactId>
+        </dependency>
+      </dependencies>
+      <build>
+        <plugins>
+          <plugin>
+            <artifactId>maven-dependency-plugin</artifactId>
+            <executions>
+              <execution>
+                <id>create-mrapp-generated-classpath</id>
+                <phase>generate-test-resources</phase>
+                <goals>
+                  <goal>build-classpath</goal>
+                </goals>
+                <configuration>
+                  <!-- needed to run the unit test for DS to generate
                                  the required classpath that is required in the env
                                  of the launch container in the mini mr/yarn cluster
                                  -->
-                                 <outputFile>${project.build.directory}/test-classes/mrapp-generated-classpath</outputFile>
-                             </configuration>
-                         </execution>
-                     </executions>
-                 </plugin>
-             </plugins>
-         </build>
-     </profile>
-     <!--
+                  <outputFile>${project.build.directory}/test-classes/mrapp-generated-classpath</outputFile>
+                </configuration>
+              </execution>
+            </executions>
+          </plugin>
+        </plugins>
+      </build>
+    </profile>
+    <!--
        profile for building against Hadoop 3.0.x. Activate using:
         mvn -Dhadoop.profile=3.0
      -->
-     <profile>
-         <id>hadoop-3.0</id>
-         <activation>
-             <property>
-                 <name>hadoop.profile</name>
-                 <value>3.0</value>
-             </property>
-         </activation>
-         <properties>
-             <hadoop.version>3.0-SNAPSHOT</hadoop.version>
-         </properties>
-         <dependencies>
-             <dependency>
-                 <groupId>org.apache.hadoop</groupId>
-                 <artifactId>hadoop-common</artifactId>
-             </dependency>
-             <dependency>
-                 <groupId>org.apache.hadoop</groupId>
-                 <artifactId>hadoop-annotations</artifactId>
-                 <exclusions>
-                   <exclusion>
-                     <groupId>jdk.tools</groupId>
-                     <artifactId>jdk.tools</artifactId>
-                   </exclusion>
-                 </exclusions>
-             </dependency>
-             <dependency>
-                 <groupId>org.apache.hadoop</groupId>
-                 <artifactId>hadoop-minicluster</artifactId>
-             </dependency>
-         </dependencies>
-         <build>
-             <plugins>
-                 <plugin>
-                     <artifactId>maven-dependency-plugin</artifactId>
-                     <executions>
-                         <execution>
-                             <id>create-mrapp-generated-classpath</id>
-                             <phase>generate-test-resources</phase>
-                             <goals>
-                                 <goal>build-classpath</goal>
-                             </goals>
-                             <configuration>
-                                 <!-- needed to run the unit test for DS to generate
+    <profile>
+      <id>hadoop-3.0</id>
+      <activation>
+        <property>
+          <name>hadoop.profile</name>
+          <value>3.0</value>
+        </property>
+      </activation>
+      <properties>
+        <hadoop.version>3.0-SNAPSHOT</hadoop.version>
+      </properties>
+      <dependencies>
+        <dependency>
+          <groupId>org.apache.hadoop</groupId>
+          <artifactId>hadoop-common</artifactId>
+        </dependency>
+        <dependency>
+          <groupId>org.apache.hadoop</groupId>
+          <artifactId>hadoop-annotations</artifactId>
+          <exclusions>
+            <exclusion>
+              <groupId>jdk.tools</groupId>
+              <artifactId>jdk.tools</artifactId>
+            </exclusion>
+          </exclusions>
+        </dependency>
+        <dependency>
+          <groupId>org.apache.hadoop</groupId>
+          <artifactId>hadoop-minicluster</artifactId>
+        </dependency>
+      </dependencies>
+      <build>
+        <plugins>
+          <plugin>
+            <artifactId>maven-dependency-plugin</artifactId>
+            <executions>
+              <execution>
+                <id>create-mrapp-generated-classpath</id>
+                <phase>generate-test-resources</phase>
+                <goals>
+                  <goal>build-classpath</goal>
+                </goals>
+                <configuration>
+                  <!-- needed to run the unit test for DS to generate
                                  the required classpath that is required in the env
                                  of the launch container in the mini mr/yarn cluster
                                  -->
-                                 <outputFile>${project.build.directory}/test-classes/mrapp-generated-classpath</outputFile>
-                             </configuration>
-                         </execution>
-                     </executions>
-                 </plugin>
-             </plugins>
-         </build>
-     </profile>
+                  <outputFile>${project.build.directory}/test-classes/mrapp-generated-classpath</outputFile>
+                </configuration>
+              </execution>
+            </executions>
+          </plugin>
+        </plugins>
+      </build>
+    </profile>
   </profiles>
 </project>

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/ExampleProtos.java
----------------------------------------------------------------------
diff --git a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/ExampleProtos.java b/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/ExampleProtos.java
deleted file mode 100644
index b780985..0000000
--- a/hbase-examples/src/main/java/org/apache/hadoop/hbase/coprocessor/example/generated/ExampleProtos.java
+++ /dev/null
@@ -1,1149 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: Examples.proto
-
-package org.apache.hadoop.hbase.coprocessor.example.generated;
-
-public final class ExampleProtos {
-  private ExampleProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  public interface CountRequestOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-  }
-  /**
-   * Protobuf type {@code hbase.pb.CountRequest}
-   */
-  public static final class CountRequest extends
-      com.google.protobuf.GeneratedMessage
-      implements CountRequestOrBuilder {
-    // Use CountRequest.newBuilder() to construct.
-    private CountRequest(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private CountRequest(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final CountRequest defaultInstance;
-    public static CountRequest getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public CountRequest getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private CountRequest(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.class, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<CountRequest> PARSER =
-        new com.google.protobuf.AbstractParser<CountRequest>() {
-      public CountRequest parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new CountRequest(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<CountRequest> getParserForType() {
-      return PARSER;
-    }
-
-    private void initFields() {
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest other = (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest) obj;
-
-      boolean result = true;
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.CountRequest}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequestOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.class, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountRequest_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest build() {
-        org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest result = new org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest(this);
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.getDefaultInstance()) return this;
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.CountRequest)
-    }
-
-    static {
-      defaultInstance = new CountRequest(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.CountRequest)
-  }
-
-  public interface CountResponseOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
-
-    // required int64 count = 1 [default = 0];
-    /**
-     * <code>required int64 count = 1 [default = 0];</code>
-     */
-    boolean hasCount();
-    /**
-     * <code>required int64 count = 1 [default = 0];</code>
-     */
-    long getCount();
-  }
-  /**
-   * Protobuf type {@code hbase.pb.CountResponse}
-   */
-  public static final class CountResponse extends
-      com.google.protobuf.GeneratedMessage
-      implements CountResponseOrBuilder {
-    // Use CountResponse.newBuilder() to construct.
-    private CountResponse(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
-      super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private CountResponse(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final CountResponse defaultInstance;
-    public static CountResponse getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public CountResponse getDefaultInstanceForType() {
-      return defaultInstance;
-    }
-
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
-    @java.lang.Override
-    public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
-      return this.unknownFields;
-    }
-    private CountResponse(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
-      int mutable_bitField0_ = 0;
-      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder();
-      try {
-        boolean done = false;
-        while (!done) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              done = true;
-              break;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                done = true;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              count_ = input.readInt64();
-              break;
-            }
-          }
-        }
-      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(this);
-      } catch (java.io.IOException e) {
-        throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
-      } finally {
-        this.unknownFields = unknownFields.build();
-        makeExtensionsImmutable();
-      }
-    }
-    public static final com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_descriptor;
-    }
-
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.Builder.class);
-    }
-
-    public static com.google.protobuf.Parser<CountResponse> PARSER =
-        new com.google.protobuf.AbstractParser<CountResponse>() {
-      public CountResponse parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new CountResponse(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<CountResponse> getParserForType() {
-      return PARSER;
-    }
-
-    private int bitField0_;
-    // required int64 count = 1 [default = 0];
-    public static final int COUNT_FIELD_NUMBER = 1;
-    private long count_;
-    /**
-     * <code>required int64 count = 1 [default = 0];</code>
-     */
-    public boolean hasCount() {
-      return ((bitField0_ & 0x00000001) == 0x00000001);
-    }
-    /**
-     * <code>required int64 count = 1 [default = 0];</code>
-     */
-    public long getCount() {
-      return count_;
-    }
-
-    private void initFields() {
-      count_ = 0L;
-    }
-    private byte memoizedIsInitialized = -1;
-    public final boolean isInitialized() {
-      byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
-
-      if (!hasCount()) {
-        memoizedIsInitialized = 0;
-        return false;
-      }
-      memoizedIsInitialized = 1;
-      return true;
-    }
-
-    public void writeTo(com.google.protobuf.CodedOutputStream output)
-                        throws java.io.IOException {
-      getSerializedSize();
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeInt64(1, count_);
-      }
-      getUnknownFields().writeTo(output);
-    }
-
-    private int memoizedSerializedSize = -1;
-    public int getSerializedSize() {
-      int size = memoizedSerializedSize;
-      if (size != -1) return size;
-
-      size = 0;
-      if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeInt64Size(1, count_);
-      }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
-      return size;
-    }
-
-    private static final long serialVersionUID = 0L;
-    @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
-    public boolean equals(final java.lang.Object obj) {
-      if (obj == this) {
-       return true;
-      }
-      if (!(obj instanceof org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse)) {
-        return super.equals(obj);
-      }
-      org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse other = (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) obj;
-
-      boolean result = true;
-      result = result && (hasCount() == other.hasCount());
-      if (hasCount()) {
-        result = result && (getCount()
-            == other.getCount());
-      }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
-      return result;
-    }
-
-    private int memoizedHashCode = 0;
-    @java.lang.Override
-    public int hashCode() {
-      if (memoizedHashCode != 0) {
-        return memoizedHashCode;
-      }
-      int hash = 41;
-      hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasCount()) {
-        hash = (37 * hash) + COUNT_FIELD_NUMBER;
-        hash = (53 * hash) + hashLong(getCount());
-      }
-      hash = (29 * hash) + getUnknownFields().hashCode();
-      memoizedHashCode = hash;
-      return hash;
-    }
-
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(
-        com.google.protobuf.ByteString data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(
-        com.google.protobuf.ByteString data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(byte[] data)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(
-        byte[] data,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws com.google.protobuf.InvalidProtocolBufferException {
-      return PARSER.parseFrom(data, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseDelimitedFrom(java.io.InputStream input)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseDelimitedFrom(
-        java.io.InputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(
-        com.google.protobuf.CodedInputStream input)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input);
-    }
-    public static org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parseFrom(
-        com.google.protobuf.CodedInputStream input,
-        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
-    }
-
-    public static Builder newBuilder() { return Builder.create(); }
-    public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse prototype) {
-      return newBuilder().mergeFrom(prototype);
-    }
-    public Builder toBuilder() { return newBuilder(this); }
-
-    @java.lang.Override
-    protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-      Builder builder = new Builder(parent);
-      return builder;
-    }
-    /**
-     * Protobuf type {@code hbase.pb.CountResponse}
-     */
-    public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponseOrBuilder {
-      public static final com.google.protobuf.Descriptors.Descriptor
-          getDescriptor() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_descriptor;
-      }
-
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
-          internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_fieldAccessorTable
-            .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.class, org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.Builder.class);
-      }
-
-      // Construct using org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.newBuilder()
-      private Builder() {
-        maybeForceBuilderInitialization();
-      }
-
-      private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
-        super(parent);
-        maybeForceBuilderInitialization();
-      }
-      private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-        }
-      }
-      private static Builder create() {
-        return new Builder();
-      }
-
-      public Builder clear() {
-        super.clear();
-        count_ = 0L;
-        bitField0_ = (bitField0_ & ~0x00000001);
-        return this;
-      }
-
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
-      public com.google.protobuf.Descriptors.Descriptor
-          getDescriptorForType() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.internal_static_hbase_pb_CountResponse_descriptor;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance();
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse build() {
-        org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(result);
-        }
-        return result;
-      }
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse buildPartial() {
-        org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse result = new org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse(this);
-        int from_bitField0_ = bitField0_;
-        int to_bitField0_ = 0;
-        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
-          to_bitField0_ |= 0x00000001;
-        }
-        result.count_ = count_;
-        result.bitField0_ = to_bitField0_;
-        onBuilt();
-        return result;
-      }
-
-      public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) {
-          return mergeFrom((org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse)other);
-        } else {
-          super.mergeFrom(other);
-          return this;
-        }
-      }
-
-      public Builder mergeFrom(org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse other) {
-        if (other == org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance()) return this;
-        if (other.hasCount()) {
-          setCount(other.getCount());
-        }
-        this.mergeUnknownFields(other.getUnknownFields());
-        return this;
-      }
-
-      public final boolean isInitialized() {
-        if (!hasCount()) {
-          
-          return false;
-        }
-        return true;
-      }
-
-      public Builder mergeFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws java.io.IOException {
-        org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse parsedMessage = null;
-        try {
-          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) e.getUnfinishedMessage();
-          throw e;
-        } finally {
-          if (parsedMessage != null) {
-            mergeFrom(parsedMessage);
-          }
-        }
-        return this;
-      }
-      private int bitField0_;
-
-      // required int64 count = 1 [default = 0];
-      private long count_ ;
-      /**
-       * <code>required int64 count = 1 [default = 0];</code>
-       */
-      public boolean hasCount() {
-        return ((bitField0_ & 0x00000001) == 0x00000001);
-      }
-      /**
-       * <code>required int64 count = 1 [default = 0];</code>
-       */
-      public long getCount() {
-        return count_;
-      }
-      /**
-       * <code>required int64 count = 1 [default = 0];</code>
-       */
-      public Builder setCount(long value) {
-        bitField0_ |= 0x00000001;
-        count_ = value;
-        onChanged();
-        return this;
-      }
-      /**
-       * <code>required int64 count = 1 [default = 0];</code>
-       */
-      public Builder clearCount() {
-        bitField0_ = (bitField0_ & ~0x00000001);
-        count_ = 0L;
-        onChanged();
-        return this;
-      }
-
-      // @@protoc_insertion_point(builder_scope:hbase.pb.CountResponse)
-    }
-
-    static {
-      defaultInstance = new CountResponse(true);
-      defaultInstance.initFields();
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.CountResponse)
-  }
-
-  /**
-   * Protobuf service {@code hbase.pb.RowCountService}
-   */
-  public static abstract class RowCountService
-      implements com.google.protobuf.Service {
-    protected RowCountService() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc getRowCount(.hbase.pb.CountRequest) returns (.hbase.pb.CountResponse);</code>
-       */
-      public abstract void getRowCount(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse> done);
-
-      /**
-       * <code>rpc getKeyValueCount(.hbase.pb.CountRequest) returns (.hbase.pb.CountResponse);</code>
-       */
-      public abstract void getKeyValueCount(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse> done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new RowCountService() {
-        @java.lang.Override
-        public  void getRowCount(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse> done) {
-          impl.getRowCount(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void getKeyValueCount(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse> done) {
-          impl.getKeyValueCount(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.getRowCount(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)request);
-            case 1:
-              return impl.getKeyValueCount(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc getRowCount(.hbase.pb.CountRequest) returns (.hbase.pb.CountResponse);</code>
-     */
-    public abstract void getRowCount(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse> done);
-
-    /**
-     * <code>rpc getKeyValueCount(.hbase.pb.CountRequest) returns (.hbase.pb.CountResponse);</code>
-     */
-    public abstract void getKeyValueCount(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse> done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.getRowCount(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse>specializeCallback(
-              done));
-          return;
-        case 1:
-          this.getKeyValueCount(controller, (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.RowCountService implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void getRowCount(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.class,
-            org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance()));
-      }
-
-      public  void getKeyValueCount(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse> done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.class,
-            org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getRowCount(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request)
-          throws com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getKeyValueCount(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request)
-          throws com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getRowCount(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse getKeyValueCount(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountRequest request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse) channel.callBlockingMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.coprocessor.example.generated.ExampleProtos.CountResponse.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:hbase.pb.RowCountService)
-  }
-
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_hbase_pb_CountRequest_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_hbase_pb_CountRequest_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
-    internal_static_hbase_pb_CountResponse_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
-      internal_static_hbase_pb_CountResponse_fieldAccessorTable;
-
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\016Examples.proto\022\010hbase.pb\"\016\n\014CountReque" +
-      "st\"!\n\rCountResponse\022\020\n\005count\030\001 \002(\003:\00102\226\001" +
-      "\n\017RowCountService\022>\n\013getRowCount\022\026.hbase" +
-      ".pb.CountRequest\032\027.hbase.pb.CountRespons" +
-      "e\022C\n\020getKeyValueCount\022\026.hbase.pb.CountRe" +
-      "quest\032\027.hbase.pb.CountResponseBN\n5org.ap" +
-      "ache.hadoop.hbase.coprocessor.example.ge" +
-      "neratedB\rExampleProtosH\001\210\001\001\240\001\001"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_hbase_pb_CountRequest_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_hbase_pb_CountRequest_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_pb_CountRequest_descriptor,
-              new java.lang.String[] { });
-          internal_static_hbase_pb_CountResponse_descriptor =
-            getDescriptor().getMessageTypes().get(1);
-          internal_static_hbase_pb_CountResponse_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_pb_CountResponse_descriptor,
-              new java.lang.String[] { "Count", });
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-procedure/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-procedure/pom.xml b/hbase-procedure/pom.xml
index 4739fa4..c3154de 100644
--- a/hbase-procedure/pom.xml
+++ b/hbase-procedure/pom.xml
@@ -70,6 +70,10 @@
       <scope>test</scope>
     </dependency>
     <dependency>
+      <groupId>org.apache.hbase.thirdparty</groupId>
+      <artifactId>hbase-shaded-protobuf</artifactId>
+    </dependency>
+    <dependency>
       <groupId>org.apache.hbase</groupId>
       <artifactId>hbase-protocol-shaded</artifactId>
     </dependency>

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/README.txt
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/README.txt b/hbase-protocol-shaded/README.txt
index 26a9665..b0030fa 100644
--- a/hbase-protocol-shaded/README.txt
+++ b/hbase-protocol-shaded/README.txt
@@ -1,53 +1,6 @@
-Please read carefully as the 'menu options' have changed.
-What you do in here is not what you do elsewhere to generate
-proto java files.
-
 This module has proto files used by core. These protos
 overlap with protos that are used by coprocessor endpoints
-(CPEP) in the module hbase-protocol. So the core versions have
+(CPEP) in the module hbase-protocol. So core versions have
 a different name, the generated classes are relocated
 -- i.e. shaded -- to a new location; they are moved from
 org.apache.hadoop.hbase.* to org.apache.hadoop.hbase.shaded.
-
-This module also includes the protobuf that hbase core depends
-on again relocated to live at an offset of
-org.apache.hadoop.hbase.shaded so as to avoid clashes with other
-versions of protobuf resident on our CLASSPATH included,
-transitively or otherwise, by dependencies: i.e. the shaded
-protobuf Message class is at
-org.apache.hadoop.hbase.shaded.com.google.protobuf.Message
-rather than at com.google.protobuf.Message.
-
-Finally, this module also includes patches applied on top of
-protobuf to add functionality not yet in protobuf that we
-need now.
-
-If you make changes to protos, to the protobuf version or to
-the patches you want to apply to protobuf, you must rerun the
-below step and then check in what it generated:
-
- $ mvn install -Dcompile-protobuf
-
-or
-
- $ mvn install -Pcompile-protobuf
-
-NOTE: 'install' above whereas other proto generation only needs 'compile'
-NOTE: Unlike elsehwere the above command does NOT install this modules jar
-into the repo., intentionally. The jar made by the above is a scratch jar
-that is part of the process that gets us to a set of files to check in;
-it is not for consumption. Run mvn install without the '-Pcompile-protobuf'
-option to get this modules' artifact installed in your repo!
-
-When finished, the content of src/main/java/org/apache/hadoop/hbase/shaded
-will have been updated. Make sure all builds and then carefully
-check in the changes. Files may have been added or removed
-by the steps above.
-
-The protobuf version used internally by hbase differs from what
-is used over in the CPEP hbase-protocol module but mvn takes care
-of ensuring we have the right protobuf in place so you don't have to.
-
-If you have patches for the protobuf, add them to
-src/main/patches directory. They will be applied after
-protobuf is shaded and unbundled into src/main/java.

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/pom.xml b/hbase-protocol-shaded/pom.xml
index 43f8ac3..fff02f8 100644
--- a/hbase-protocol-shaded/pom.xml
+++ b/hbase-protocol-shaded/pom.xml
@@ -32,18 +32,12 @@
   <properties>
     <maven.javadoc.skip>true</maven.javadoc.skip>
     <!--Version of protobuf that hbase uses internally (we shade our pb)
+         Must match what is out in hbase-thirdparty include.
+         (Note, there may only be a 3.3.0 protoc... no 3.3.1 protoc
            -->
-    <internal.protobuf.version>3.2.0</internal.protobuf.version>
-    <!--The Default target dir-->
-    <classes.dir>${project.build.directory}/classes</classes.dir>
-    <!--The Default location for sources-->
-    <sources.dir>src/main/java</sources.dir>
+    <internal.protobuf.version>3.3.0</internal.protobuf.version>
   </properties>
   <build>
-    <!--I want to override these in profile so define them
-         with variables up here-->
-    <sourceDirectory>${sources.dir}</sourceDirectory>
-    <outputDirectory>${classes.dir}</outputDirectory>
     <plugins>
       <plugin>
         <groupId>org.apache.maven.plugins</groupId>
@@ -56,7 +50,7 @@
       <plugin>
         <groupId>org.apache.maven.plugins</groupId>
         <artifactId>maven-source-plugin</artifactId>
-      </plugin>
+		  </plugin> 
       <plugin>
         <!--Make it so assembly:single does nothing in here-->
         <artifactId>maven-assembly-plugin</artifactId>
@@ -80,6 +74,101 @@
           </execution>
         </executions>
       </plugin>
+      <plugin>
+        <groupId>org.xolstice.maven.plugins</groupId>
+        <artifactId>protobuf-maven-plugin</artifactId>
+        <executions>
+          <execution>
+            <id>compile-protoc</id>
+            <phase>generate-sources</phase>
+            <goals>
+              <goal>compile</goal>
+            </goals>
+            <configuration>
+              <protocArtifact>com.google.protobuf:protoc:${internal.protobuf.version}:exe:${os.detected.classifier}</protocArtifact>
+              <attachProtoSources>false</attachProtoSources>
+              <checkStaleness>true</checkStaleness>
+            </configuration>
+          </execution>
+        </executions>
+      </plugin>
+      <!--Need this old plugin to replace in generated files instances
+           of com.google.protobuf so instead its o.a.h.h.com.google.protobuf.
+           Plugin is old and in google code archive. Here is usage done by
+           anohther: https://github.com/beiliubei/maven-replacer-plugin/wiki/Usage-Guide
+           The mess with the regex in the below is to prevent replacement every time
+           we run mvn install. There is probably a better way of avoiding the
+           double interpolation but this is it for now.
+        -->
+      <plugin> 
+        <groupId>com.google.code.maven-replacer-plugin</groupId> 
+        <artifactId>replacer</artifactId> 
+        <version>1.5.3</version> 
+        <executions> 
+          <execution> 
+            <phase>generate-sources</phase>
+            <goals> 
+              <goal>replace</goal>
+            </goals> 
+          </execution> 
+        </executions> 
+        <configuration> 
+          <basedir>${basedir}/target/generated-sources/</basedir>  
+            <includes>  
+                <include>**/*.java</include>  
+            </includes>  
+            <replacements>
+              <replacement>
+                <token>([^\.])com.google.protobuf</token> 
+                <value>$1org.apache.hadoop.hbase.shaded.com.google.protobuf</value> 
+              </replacement>
+            </replacements>
+        </configuration> 
+      </plugin>
+      <plugin>
+        <groupId>org.apache.maven.plugins</groupId>
+        <artifactId>maven-shade-plugin</artifactId>
+        <executions>
+          <execution>
+            <phase>package</phase>
+            <goals>
+              <goal>shade</goal>
+            </goals>
+            <configuration>
+              <minimizeJar>true</minimizeJar>
+              <shadeSourcesContent>true</shadeSourcesContent>
+              <!-- Causes an NPE until shade 3.0.1.  See MSHADE-247
+              <createSourcesJar>true</createSourcesJar>
+              -->
+              <relocations>
+                <relocation>
+                  <pattern>com.google.protobuf</pattern>
+                  <shadedPattern>org.apache.hadoop.hbase.shaded.com.google.protobuf</shadedPattern>
+                </relocation>
+              </relocations>
+              <artifactSet>
+                <excludes>
+                  <!--Exclude protobuf itself. We get a patched version from hbase-thirdparty.
+                    -->
+                  <exclude>org.apache.hadoop.hbase.shaded.com.google:*</exclude>
+                  <exclude>com.google.protobuf:protobuf-java</exclude>
+                  <exclude>com.google.code.findbugs:*</exclude>
+                  <exclude>com.google.errorprone:error_prone_annotations</exclude>
+                  <exclude>com.google.j2objc:j2objc-annotations</exclude>
+                  <exclude>org.codehaus.mojo:animal-sniffer-annotations</exclude>
+                  <exclude>org.codehaus.mojo:animal-sniffer-annotations</exclude>
+                  <exclude>junit:junit</exclude>
+                  <exclude>log4j:log4j</exclude>
+                  <exclude>commons-logging:commons-logging</exclude>
+                  <exclude>org.apache.hbase:hbase-annotations</exclude>
+                  <exclude>com.github.stephenc.fingbugs:*</exclude>
+                  <exclude></exclude>
+                </excludes>
+              </artifactSet>
+            </configuration>
+          </execution>
+        </executions>
+      </plugin>
     </plugins>
     <pluginManagement>
       <plugins>
@@ -116,6 +205,10 @@
           will get bundled-->
     <!-- Intra-project dependencies -->
     <dependency>
+      <groupId>org.apache.hbase.thirdparty</groupId>
+      <artifactId>hbase-shaded-protobuf</artifactId>
+    </dependency>
+    <dependency>
       <groupId>org.apache.hbase</groupId>
       <artifactId>hbase-annotations</artifactId>
       <exclusions>
@@ -147,218 +240,8 @@
       </activation>
       <properties>
         <surefire.skipFirstPart>true</surefire.skipFirstPart>
+        <surefire.skipSecondPart>true</surefire.skipSecondPart>
       </properties>
     </profile>
-    <profile>
-      <id>compile-protobuf</id>
-      <!--
-         Generate and shade proto files. Drops generated java files
-         under src/main/java when done. Check in the generated files so
-         available at build time. Run this profile/step everytime you change
-         proto files or update the protobuf version.
-
-         The below does a bunch of ugly stuff. It purges current content
-         of the generated and shaded com.google.protobuf java files first.
-         Let me say that again. We do a remove of java files under src/main/java
-         in the shaded dirs. It does this because later we apply patches and
-         patches fail if they've already been applied. We remove too because we
-         overlay the shaded protobuf and if files have been removed or added,
-         it'll be more plain if we have first done this delete.
-
-         Next up we generate protos, build a scratch jar that contains protos
-         only and stuff we want shaded (we have to do this because shading only
-         works at install time on a jar), run the shade on the jar, then
-         carefully STOP this scratch jar from being put into the local repository
-         (because it can mess up builds that come later... mvn automatically wants
-         to install artifact into repo per module). Finally, undo this shaded
-         jar over the src/main/java directory, and then apply patches atop this.
-
-         The result needs to be checked in.
-      -->
-      <activation>
-        <property>
-          <name>compile-protobuf</name>
-        </property>
-      </activation>
-      <properties>
-        <profile.id>compile-protobuf</profile.id>
-        <!--Directory under target to hold generated protos files-->
-        <protoc.sources.dir>${project.build.directory}/protoc-generated-sources</protoc.sources.dir>
-        <!--When doing this step, the sources.dir is pointed at generated protos, NOT src/main/java-->
-        <sources.dir>${protoc.sources.dir}</sources.dir>
-        <!--Where to compile protos into-->
-        <classes.dir>${project.build.directory}/protoc-generated-classes</classes.dir>
-        <!--When the compile for this profile runs, make sure it makes jars that
-            can be related back to this shading profile. Give them the shading profile
-            name as a prefix.
-         -->
-        <jar.finalName>${profile.id}.${project.artifactId}-${project.version}</jar.finalName>
-      </properties>
-      <build>
-        <plugins>
-          <plugin>
-            <artifactId>maven-clean-plugin</artifactId>
-              <executions>
-                <execution>
-                  <id>pre-compile-protoc</id>
-                  <phase>generate-sources</phase>
-                  <goals>
-                    <goal>clean</goal>
-                  </goals>
-                <configuration>
-                  <filesets>
-                    <fileset>
-                      <directory>${basedir}/src/main/java/org/apache/hadoop/hbase/shaded</directory>
-                      <includes>
-                        <include>ipc/protobuf/generated/**/*.java</include>
-                        <include>protobuf/generated/**/*.java</include>
-                        <include>com/google/protobuf/**/*.java</include>
-                      </includes>
-                      <followSymlinks>false</followSymlinks>
-                    </fileset>
-                  </filesets>
-                </configuration>
-              </execution>
-            </executions>
-          </plugin>
-          <plugin>
-            <groupId>org.xolstice.maven.plugins</groupId>
-            <artifactId>protobuf-maven-plugin</artifactId>
-            <executions>
-              <execution>
-                <id>compile-protoc</id>
-                <phase>generate-sources</phase>
-                <goals>
-                  <goal>compile</goal>
-                </goals>
-                <configuration>
-                  <protocArtifact>com.google.protobuf:protoc:${internal.protobuf.version}:exe:${os.detected.classifier}</protocArtifact>
-                  <outputDirectory>${protoc.sources.dir}</outputDirectory>
-                  <attachProtoSources>false</attachProtoSources>
-                </configuration>
-              </execution>
-            </executions>
-          </plugin>
-          <plugin>
-            <groupId>org.apache.maven.plugins</groupId>
-            <artifactId>maven-jar-plugin</artifactId>
-            <version>${maven.jar.version}</version>
-            <configuration>
-              <finalName>${jar.finalName}</finalName>                   
-            </configuration>
-          </plugin>
-          <plugin>
-            <groupId>org.apache.maven.plugins</groupId>
-            <artifactId>maven-shade-plugin</artifactId>
-            <version>${maven.shade.version}</version>
-            <executions>
-              <execution>
-                <phase>package</phase>
-                <goals>
-                  <goal>shade</goal>
-                </goals>
-                <configuration>
-                  <shadeSourcesContent>true</shadeSourcesContent>
-                  <createSourcesJar>true</createSourcesJar>
-                  <relocations>
-                    <relocation>
-                      <pattern>com.google.protobuf</pattern>
-                      <shadedPattern>org.apache.hadoop.hbase.shaded.com.google.protobuf</shadedPattern>
-                    </relocation>
-                  </relocations>
-                  <!-- What I got when I did a mvn dependency:list for this
-                   module. Exclude all but the protobuf
-                [INFO]    commons-logging:commons-logging:jar:1.2:compile
-                [INFO]    com.github.stephenc.findbugs:findbugs-annotations:jar:1.3.9-1:compile
-                [INFO]    log4j:log4j:jar:1.2.17:compile
-                [INFO]    com.google.protobuf:protobuf-java:jar:2.5.0:compile
-                [INFO]    org.hamcrest:hamcrest-core:jar:1.3:test
-                [INFO]    org.mockito:mockito-all:jar:1.10.8:test
-                [INFO]    junit:junit:jar:4.12:compile
-                [INFO]    org.apache.hbase:hbase-annotations:jar:2.0.0-SNAPSHOT:compile
-
-                  The list below must exlude all of the above except protobuf.
-              -->
-                  <artifactSet>
-                    <excludes>
-                      <exclude>commons-logging:commons-logging</exclude>
-                      <exclude>com.github.stephenc.findbugs:findbugs-annotations</exclude>
-                      <exclude>log4j:log4j</exclude>
-                      <exclude>org.hamcrest:hamcrest-core</exclude>
-                      <exclude>org.mockito:mockito-all</exclude>
-                      <exclude>junit:junit</exclude>
-                      <exclude>org.apache.hbase:hbase-annotations</exclude>
-                    </excludes>
-                  </artifactSet>
-                </configuration>
-              </execution>
-            </executions>
-          </plugin>
-          <!--Now unpack the shaded jar made above so the shaded classes
-             are available to subsequent modules-->
-          <plugin>
-            <groupId>org.apache.maven.plugins</groupId>
-            <artifactId>maven-dependency-plugin</artifactId>
-            <executions>
-              <execution>
-                <id>unpack</id>
-                <phase>package</phase>
-                <goals>
-                  <goal>unpack</goal>
-                </goals>
-                <configuration>
-                  <artifactItems>
-                    <artifactItem>
-                      <groupId>${project.groupId}</groupId>
-                      <artifactId>${project.artifactId}</artifactId>
-                      <version>${project.version}</version>
-                      <classifier>sources</classifier>
-                      <type>jar</type>
-                      <overWrite>true</overWrite>
-                      <outputDirectory>${basedir}/src/main/java</outputDirectory>
-                      <includes>**/*.java</includes>
-                    </artifactItem>
-                  </artifactItems>
-                </configuration>
-              </execution>
-            </executions>
-          </plugin>
-      <plugin>
-        <groupId>org.apache.maven.plugins</groupId>
-        <artifactId>maven-patch-plugin</artifactId>
-        <version>${maven.patch.version}</version>
-        <configuration>
-           <!--Patches are made at top-level-->
-           <targetDirectory>${basedir}/..</targetDirectory>
-          <skipApplication>false</skipApplication>
-        </configuration>
-        <executions>
-          <execution>
-            <id>patch</id>
-            <configuration>
-               <strip>1</strip>
-              <patchDirectory>src/main/patches</patchDirectory>
-              <patchTrackingFile>${project.build.directory}/patches-applied.txt</patchTrackingFile>
-              <naturalOrderProcessing>true</naturalOrderProcessing>
-            </configuration>
-            <phase>package</phase>
-            <goals>
-              <!--This should run after the above unpack phase-->
-              <goal>apply</goal>
-            </goals>
-          </execution>
-        </executions>
-      </plugin>
-        <plugin>
-            <groupId>org.apache.maven.plugins</groupId>
-            <artifactId>maven-install-plugin</artifactId>
-            <version>${maven.install.version}</version>
-            <configuration>
-                <skip>true</skip>
-            </configuration>
-        </plugin>
-        </plugins>
-      </build>
-    </profile>
   </profiles>
 </project>


[11/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt32Value.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt32Value.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt32Value.java
deleted file mode 100644
index 7b4dae9..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt32Value.java
+++ /dev/null
@@ -1,451 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `uint32`.
- * The JSON representation for `UInt32Value` is JSON number.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.UInt32Value}
- */
-public  final class UInt32Value extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.UInt32Value)
-    UInt32ValueOrBuilder {
-  // Use UInt32Value.newBuilder() to construct.
-  private UInt32Value(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private UInt32Value() {
-    value_ = 0;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private UInt32Value(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-
-            value_ = input.readUInt32();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt32Value_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt32Value_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private int value_;
-  /**
-   * <pre>
-   * The uint32 value.
-   * </pre>
-   *
-   * <code>uint32 value = 1;</code>
-   */
-  public int getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (value_ != 0) {
-      output.writeUInt32(1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (value_ != 0) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeUInt32Size(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value) obj;
-
-    boolean result = true;
-    result = result && (getValue()
-        == other.getValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + getValue();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `uint32`.
-   * The JSON representation for `UInt32Value` is JSON number.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.UInt32Value}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.UInt32Value)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32ValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt32Value_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt32Value_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = 0;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt32Value_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value.getDefaultInstance()) return this;
-      if (other.getValue() != 0) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private int value_ ;
-    /**
-     * <pre>
-     * The uint32 value.
-     * </pre>
-     *
-     * <code>uint32 value = 1;</code>
-     */
-    public int getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * The uint32 value.
-     * </pre>
-     *
-     * <code>uint32 value = 1;</code>
-     */
-    public Builder setValue(int value) {
-      
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The uint32 value.
-     * </pre>
-     *
-     * <code>uint32 value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = 0;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.UInt32Value)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.UInt32Value)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UInt32Value>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<UInt32Value>() {
-    public UInt32Value parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new UInt32Value(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UInt32Value> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UInt32Value> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt32Value getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt32ValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt32ValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt32ValueOrBuilder.java
deleted file mode 100644
index ad0d537..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt32ValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface UInt32ValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.UInt32Value)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The uint32 value.
-   * </pre>
-   *
-   * <code>uint32 value = 1;</code>
-   */
-  int getValue();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt64Value.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt64Value.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt64Value.java
deleted file mode 100644
index 578a6ff..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt64Value.java
+++ /dev/null
@@ -1,452 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * Wrapper message for `uint64`.
- * The JSON representation for `UInt64Value` is JSON string.
- * </pre>
- *
- * Protobuf type {@code google.protobuf.UInt64Value}
- */
-public  final class UInt64Value extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.UInt64Value)
-    UInt64ValueOrBuilder {
-  // Use UInt64Value.newBuilder() to construct.
-  private UInt64Value(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private UInt64Value() {
-    value_ = 0L;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private UInt64Value(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 8: {
-
-            value_ = input.readUInt64();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt64Value_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt64Value_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value.Builder.class);
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 1;
-  private long value_;
-  /**
-   * <pre>
-   * The uint64 value.
-   * </pre>
-   *
-   * <code>uint64 value = 1;</code>
-   */
-  public long getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (value_ != 0L) {
-      output.writeUInt64(1, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (value_ != 0L) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeUInt64Size(1, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value) obj;
-
-    boolean result = true;
-    result = result && (getValue()
-        == other.getValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashLong(
-        getValue());
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * Wrapper message for `uint64`.
-   * The JSON representation for `UInt64Value` is JSON string.
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.UInt64Value}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.UInt64Value)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64ValueOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt64Value_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt64Value_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      value_ = 0L;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_UInt64Value_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value(this);
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value.getDefaultInstance()) return this;
-      if (other.getValue() != 0L) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private long value_ ;
-    /**
-     * <pre>
-     * The uint64 value.
-     * </pre>
-     *
-     * <code>uint64 value = 1;</code>
-     */
-    public long getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * The uint64 value.
-     * </pre>
-     *
-     * <code>uint64 value = 1;</code>
-     */
-    public Builder setValue(long value) {
-      
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * The uint64 value.
-     * </pre>
-     *
-     * <code>uint64 value = 1;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = 0L;
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.UInt64Value)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.UInt64Value)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UInt64Value>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<UInt64Value>() {
-    public UInt64Value parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new UInt64Value(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UInt64Value> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<UInt64Value> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.UInt64Value getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt64ValueOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt64ValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt64ValueOrBuilder.java
deleted file mode 100644
index 5f65674..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UInt64ValueOrBuilder.java
+++ /dev/null
@@ -1,18 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/wrappers.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface UInt64ValueOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.UInt64Value)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * The uint64 value.
-   * </pre>
-   *
-   * <code>uint64 value = 1;</code>
-   */
-  long getValue();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UninitializedMessageException.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UninitializedMessageException.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UninitializedMessageException.java
deleted file mode 100644
index ede9d0d..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UninitializedMessageException.java
+++ /dev/null
@@ -1,99 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import java.util.Collections;
-import java.util.List;
-
-/**
- * Thrown when attempting to build a protocol message that is missing required
- * fields.  This is a {@code RuntimeException} because it normally represents
- * a programming error:  it happens when some code which constructs a message
- * fails to set all the fields.  {@code parseFrom()} methods <b>do not</b>
- * throw this; they throw an {@link InvalidProtocolBufferException} if
- * required fields are missing, because it is not a programming error to
- * receive an incomplete message.  In other words,
- * {@code UninitializedMessageException} should never be thrown by correct
- * code, but {@code InvalidProtocolBufferException} might be.
- *
- * @author kenton@google.com Kenton Varda
- */
-public class UninitializedMessageException extends RuntimeException {
-  private static final long serialVersionUID = -7466929953374883507L;
-
-  public UninitializedMessageException(final MessageLite message) {
-    super("Message was missing required fields.  (Lite runtime could not " +
-          "determine which fields were missing).");
-    missingFields = null;
-  }
-
-  public UninitializedMessageException(final List<String> missingFields) {
-    super(buildDescription(missingFields));
-    this.missingFields = missingFields;
-  }
-
-  private final List<String> missingFields;
-
-  /**
-   * Get a list of human-readable names of required fields missing from this
-   * message.  Each name is a full path to a field, e.g. "foo.bar[5].baz".
-   * Returns null if the lite runtime was used, since it lacks the ability to
-   * find missing fields.
-   */
-  public List<String> getMissingFields() {
-    return Collections.unmodifiableList(missingFields);
-  }
-
-  /**
-   * Converts this exception to an {@link InvalidProtocolBufferException}.
-   * When a parsed message is missing required fields, this should be thrown
-   * instead of {@code UninitializedMessageException}.
-   */
-  public InvalidProtocolBufferException asInvalidProtocolBufferException() {
-    return new InvalidProtocolBufferException(getMessage());
-  }
-
-  /** Construct the description string for this exception. */
-  private static String buildDescription(final List<String> missingFields) {
-    final StringBuilder description =
-      new StringBuilder("Message missing required fields: ");
-    boolean first = true;
-    for (final String field : missingFields) {
-      if (first) {
-        first = false;
-      } else {
-        description.append(", ");
-      }
-      description.append(field);
-    }
-    return description.toString();
-  }
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnknownFieldSet.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnknownFieldSet.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnknownFieldSet.java
deleted file mode 100644
index 148a3dd..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/UnknownFieldSet.java
+++ /dev/null
@@ -1,1042 +0,0 @@
-// Protocol Buffers - Google's data interchange format
-// Copyright 2008 Google Inc.  All rights reserved.
-// https://developers.google.com/protocol-buffers/
-//
-// Redistribution and use in source and binary forms, with or without
-// modification, are permitted provided that the following conditions are
-// met:
-//
-//     * Redistributions of source code must retain the above copyright
-// notice, this list of conditions and the following disclaimer.
-//     * Redistributions in binary form must reproduce the above
-// copyright notice, this list of conditions and the following disclaimer
-// in the documentation and/or other materials provided with the
-// distribution.
-//     * Neither the name of Google Inc. nor the names of its
-// contributors may be used to endorse or promote products derived from
-// this software without specific prior written permission.
-//
-// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-import org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractMessageLite.Builder.LimitedInputStream;
-import java.io.IOException;
-import java.io.InputStream;
-import java.io.OutputStream;
-import java.util.ArrayList;
-import java.util.Arrays;
-import java.util.Collections;
-import java.util.List;
-import java.util.ListIterator;
-import java.util.Map;
-import java.util.TreeMap;
-
-/**
- * {@code UnknownFieldSet} is used to keep track of fields which were seen when
- * parsing a protocol message but whose field numbers or types are unrecognized.
- * This most frequently occurs when new fields are added to a message type
- * and then messages containing those fields are read by old software that was
- * compiled before the new types were added.
- *
- * <p>Every {@link Message} contains an {@code UnknownFieldSet} (and every
- * {@link Message.Builder} contains an {@link Builder}).
- *
- * <p>Most users will never need to use this class.
- *
- * @author kenton@google.com Kenton Varda
- */
-public final class UnknownFieldSet implements MessageLite {
-
-  private UnknownFieldSet() {
-    fields = null;
-  }
-
-  /** Create a new {@link Builder}. */
-  public static Builder newBuilder() {
-    return Builder.create();
-  }
-
-  /**
-   * Create a new {@link Builder} and initialize it to be a copy
-   * of {@code copyFrom}.
-   */
-  public static Builder newBuilder(final UnknownFieldSet copyFrom) {
-    return newBuilder().mergeFrom(copyFrom);
-  }
-
-  /** Get an empty {@code UnknownFieldSet}. */
-  public static UnknownFieldSet getDefaultInstance() {
-    return defaultInstance;
-  }
-  @Override
-  public UnknownFieldSet getDefaultInstanceForType() {
-    return defaultInstance;
-  }
-  private static final UnknownFieldSet defaultInstance =
-    new UnknownFieldSet(Collections.<Integer, Field>emptyMap(),
-        Collections.<Integer, Field>emptyMap());
-
-  /**
-   * Construct an {@code UnknownFieldSet} around the given map.  The map is
-   * expected to be immutable.
-   */
-  private UnknownFieldSet(final Map<Integer, Field> fields,
-      final Map<Integer, Field> fieldsDescending) {
-    this.fields = fields;
-  }
-  private final Map<Integer, Field> fields;
-
-
-  @Override
-  public boolean equals(final Object other) {
-    if (this == other) {
-      return true;
-    }
-    return (other instanceof UnknownFieldSet) &&
-           fields.equals(((UnknownFieldSet) other).fields);
-  }
-
-  @Override
-  public int hashCode() {
-    return fields.hashCode();
-  }
-
-  /** Get a map of fields in the set by number. */
-  public Map<Integer, Field> asMap() {
-    return fields;
-  }
-
-  /** Check if the given field number is present in the set. */
-  public boolean hasField(final int number) {
-    return fields.containsKey(number);
-  }
-
-  /**
-   * Get a field by number.  Returns an empty field if not present.  Never
-   * returns {@code null}.
-   */
-  public Field getField(final int number) {
-    final Field result = fields.get(number);
-    return (result == null) ? Field.getDefaultInstance() : result;
-  }
-
-  /** Serializes the set and writes it to {@code output}. */
-  @Override
-  public void writeTo(final CodedOutputStream output) throws IOException {
-    for (final Map.Entry<Integer, Field> entry : fields.entrySet()) {
-      Field field = entry.getValue();
-      field.writeTo(entry.getKey(), output);
-    }
-  }
-
-  /**
-   * Converts the set to a string in protocol buffer text format. This is
-   * just a trivial wrapper around
-   * {@link TextFormat#printToString(UnknownFieldSet)}.
-   */
-  @Override
-  public String toString() {
-    return TextFormat.printToString(this);
-  }
-
-  /**
-   * Serializes the message to a {@code ByteString} and returns it. This is
-   * just a trivial wrapper around {@link #writeTo(CodedOutputStream)}.
-   */
-  @Override
-  public ByteString toByteString() {
-    try {
-      final ByteString.CodedBuilder out =
-        ByteString.newCodedBuilder(getSerializedSize());
-      writeTo(out.getCodedOutput());
-      return out.build();
-    } catch (final IOException e) {
-      throw new RuntimeException(
-        "Serializing to a ByteString threw an IOException (should " +
-        "never happen).", e);
-    }
-  }
-
-  /**
-   * Serializes the message to a {@code byte} array and returns it.  This is
-   * just a trivial wrapper around {@link #writeTo(CodedOutputStream)}.
-   */
-  @Override
-  public byte[] toByteArray() {
-    try {
-      final byte[] result = new byte[getSerializedSize()];
-      final CodedOutputStream output = CodedOutputStream.newInstance(result);
-      writeTo(output);
-      output.checkNoSpaceLeft();
-      return result;
-    } catch (final IOException e) {
-      throw new RuntimeException(
-        "Serializing to a byte array threw an IOException " +
-        "(should never happen).", e);
-    }
-  }
-
-  /**
-   * Serializes the message and writes it to {@code output}.  This is just a
-   * trivial wrapper around {@link #writeTo(CodedOutputStream)}.
-   */
-  @Override
-  public void writeTo(final OutputStream output) throws IOException {
-    final CodedOutputStream codedOutput = CodedOutputStream.newInstance(output);
-    writeTo(codedOutput);
-    codedOutput.flush();
-  }
-
-  @Override
-  public void writeDelimitedTo(OutputStream output) throws IOException {
-    final CodedOutputStream codedOutput = CodedOutputStream.newInstance(output);
-    codedOutput.writeRawVarint32(getSerializedSize());
-    writeTo(codedOutput);
-    codedOutput.flush();
-  }
-
-  /** Get the number of bytes required to encode this set. */
-  @Override
-  public int getSerializedSize() {
-    int result = 0;
-    for (final Map.Entry<Integer, Field> entry : fields.entrySet()) {
-      result += entry.getValue().getSerializedSize(entry.getKey());
-    }
-    return result;
-  }
-
-  /**
-   * Serializes the set and writes it to {@code output} using
-   * {@code MessageSet} wire format.
-   */
-  public void writeAsMessageSetTo(final CodedOutputStream output)
-      throws IOException {
-    for (final Map.Entry<Integer, Field> entry : fields.entrySet()) {
-      entry.getValue().writeAsMessageSetExtensionTo(
-        entry.getKey(), output);
-    }
-  }
-
-
-  /** Get the number of bytes required to encode this set using {@code MessageSet} wire format. */
-  public int getSerializedSizeAsMessageSet() {
-    int result = 0;
-    for (final Map.Entry<Integer, Field> entry : fields.entrySet()) {
-      result += entry.getValue().getSerializedSizeAsMessageSetExtension(
-        entry.getKey());
-    }
-    return result;
-  }
-
-  @Override
-  public boolean isInitialized() {
-    // UnknownFieldSets do not have required fields, so they are always
-    // initialized.
-    return true;
-  }
-
-  /** Parse an {@code UnknownFieldSet} from the given input stream. */
-  public static UnknownFieldSet parseFrom(final CodedInputStream input)
-                                          throws IOException {
-    return newBuilder().mergeFrom(input).build();
-  }
-
-  /** Parse {@code data} as an {@code UnknownFieldSet} and return it. */
-  public static UnknownFieldSet parseFrom(final ByteString data)
-      throws InvalidProtocolBufferException {
-    return newBuilder().mergeFrom(data).build();
-  }
-
-  /** Parse {@code data} as an {@code UnknownFieldSet} and return it. */
-  public static UnknownFieldSet parseFrom(final byte[] data)
-      throws InvalidProtocolBufferException {
-    return newBuilder().mergeFrom(data).build();
-  }
-
-  /** Parse an {@code UnknownFieldSet} from {@code input} and return it. */
-  public static UnknownFieldSet parseFrom(final InputStream input)
-                                          throws IOException {
-    return newBuilder().mergeFrom(input).build();
-  }
-
-  @Override
-  public Builder newBuilderForType() {
-    return newBuilder();
-  }
-
-  @Override
-  public Builder toBuilder() {
-    return newBuilder().mergeFrom(this);
-  }
-
-  /**
-   * Builder for {@link UnknownFieldSet}s.
-   *
-   * <p>Note that this class maintains {@link Field.Builder}s for all fields
-   * in the set.  Thus, adding one element to an existing {@link Field} does not
-   * require making a copy.  This is important for efficient parsing of
-   * unknown repeated fields.  However, it implies that {@link Field}s cannot
-   * be constructed independently, nor can two {@link UnknownFieldSet}s share
-   * the same {@code Field} object.
-   *
-   * <p>Use {@link UnknownFieldSet#newBuilder()} to construct a {@code Builder}.
-   */
-  public static final class Builder implements MessageLite.Builder {
-    // This constructor should never be called directly (except from 'create').
-    private Builder() {}
-
-    private Map<Integer, Field> fields;
-
-    // Optimization:  We keep around a builder for the last field that was
-    //   modified so that we can efficiently add to it multiple times in a
-    //   row (important when parsing an unknown repeated field).
-    private int lastFieldNumber;
-    private Field.Builder lastField;
-
-    private static Builder create() {
-      Builder builder = new Builder();
-      builder.reinitialize();
-      return builder;
-    }
-
-    /**
-     * Get a field builder for the given field number which includes any
-     * values that already exist.
-     */
-    private Field.Builder getFieldBuilder(final int number) {
-      if (lastField != null) {
-        if (number == lastFieldNumber) {
-          return lastField;
-        }
-        // Note:  addField() will reset lastField and lastFieldNumber.
-        addField(lastFieldNumber, lastField.build());
-      }
-      if (number == 0) {
-        return null;
-      } else {
-        final Field existing = fields.get(number);
-        lastFieldNumber = number;
-        lastField = Field.newBuilder();
-        if (existing != null) {
-          lastField.mergeFrom(existing);
-        }
-        return lastField;
-      }
-    }
-
-    /**
-     * Build the {@link UnknownFieldSet} and return it.
-     *
-     * <p>Once {@code build()} has been called, the {@code Builder} will no
-     * longer be usable.  Calling any method after {@code build()} will result
-     * in undefined behavior and can cause a {@code NullPointerException} to be
-     * thrown.
-     */
-    @Override
-    public UnknownFieldSet build() {
-      getFieldBuilder(0); // Force lastField to be built.
-      final UnknownFieldSet result;
-      if (fields.isEmpty()) {
-        result = getDefaultInstance();
-      } else {
-        Map<Integer, Field> descendingFields = null;
-        result = new UnknownFieldSet(Collections.unmodifiableMap(fields), descendingFields);
-      }
-      fields = null;
-      return result;
-    }
-
-    @Override
-    public UnknownFieldSet buildPartial() {
-      // No required fields, so this is the same as build().
-      return build();
-    }
-
-    @Override
-    public Builder clone() {
-      getFieldBuilder(0);  // Force lastField to be built.
-      Map<Integer, Field> descendingFields = null;
-      return UnknownFieldSet.newBuilder().mergeFrom(
-          new UnknownFieldSet(fields, descendingFields));
-    }
-
-    @Override
-    public UnknownFieldSet getDefaultInstanceForType() {
-      return UnknownFieldSet.getDefaultInstance();
-    }
-
-    private void reinitialize() {
-      fields = Collections.emptyMap();
-      lastFieldNumber = 0;
-      lastField = null;
-    }
-
-    /** Reset the builder to an empty set. */
-    @Override
-    public Builder clear() {
-      reinitialize();
-      return this;
-    }
-    
-    /** Clear fields from the set with a given field number. */
-    public Builder clearField(final int number) {
-      if (number == 0) {
-        throw new IllegalArgumentException("Zero is not a valid field number.");
-      }
-      if (lastField != null && lastFieldNumber == number) {
-        // Discard this.
-        lastField = null;
-        lastFieldNumber = 0;
-      }
-      if (fields.containsKey(number)) {
-        fields.remove(number);
-      }
-      return this;
-    }
-
-    /**
-     * Merge the fields from {@code other} into this set.  If a field number
-     * exists in both sets, {@code other}'s values for that field will be
-     * appended to the values in this set.
-     */
-    public Builder mergeFrom(final UnknownFieldSet other) {
-      if (other != getDefaultInstance()) {
-        for (final Map.Entry<Integer, Field> entry : other.fields.entrySet()) {
-          mergeField(entry.getKey(), entry.getValue());
-        }
-      }
-      return this;
-    }
-
-    /**
-     * Add a field to the {@code UnknownFieldSet}.  If a field with the same
-     * number already exists, the two are merged.
-     */
-    public Builder mergeField(final int number, final Field field) {
-      if (number == 0) {
-        throw new IllegalArgumentException("Zero is not a valid field number.");
-      }
-      if (hasField(number)) {
-        getFieldBuilder(number).mergeFrom(field);
-      } else {
-        // Optimization:  We could call getFieldBuilder(number).mergeFrom(field)
-        // in this case, but that would create a copy of the Field object.
-        // We'd rather reuse the one passed to us, so call addField() instead.
-        addField(number, field);
-      }
-      return this;
-    }
-
-    /**
-     * Convenience method for merging a new field containing a single varint
-     * value.  This is used in particular when an unknown enum value is
-     * encountered.
-     */
-    public Builder mergeVarintField(final int number, final int value) {
-      if (number == 0) {
-        throw new IllegalArgumentException("Zero is not a valid field number.");
-      }
-      getFieldBuilder(number).addVarint(value);
-      return this;
-    }
-
-
-    /**
-     * Convenience method for merging a length-delimited field.
-     *
-     * <p>For use by generated code only.
-     */
-    public Builder mergeLengthDelimitedField(
-        final int number, final ByteString value) {  
-      if (number == 0) {
-        throw new IllegalArgumentException("Zero is not a valid field number.");
-      }
-      getFieldBuilder(number).addLengthDelimited(value);
-      return this;
-    }
-
-    /** Check if the given field number is present in the set. */
-    public boolean hasField(final int number) {
-      if (number == 0) {
-        throw new IllegalArgumentException("Zero is not a valid field number.");
-      }
-      return number == lastFieldNumber || fields.containsKey(number);
-    }
-
-    /**
-     * Add a field to the {@code UnknownFieldSet}.  If a field with the same
-     * number already exists, it is removed.
-     */
-    public Builder addField(final int number, final Field field) {
-      if (number == 0) {
-        throw new IllegalArgumentException("Zero is not a valid field number.");
-      }
-      if (lastField != null && lastFieldNumber == number) {
-        // Discard this.
-        lastField = null;
-        lastFieldNumber = 0;
-      }
-      if (fields.isEmpty()) {
-        fields = new TreeMap<Integer,Field>();
-      }
-      fields.put(number, field);
-      return this;
-    }
-
-    /**
-     * Get all present {@code Field}s as an immutable {@code Map}.  If more
-     * fields are added, the changes may or may not be reflected in this map.
-     */
-    public Map<Integer, Field> asMap() {
-      getFieldBuilder(0);  // Force lastField to be built.
-      return Collections.unmodifiableMap(fields);
-    }
-
-    /**
-     * Parse an entire message from {@code input} and merge its fields into
-     * this set.
-     */
-    @Override
-    public Builder mergeFrom(final CodedInputStream input) throws IOException {
-      while (true) {
-        final int tag = input.readTag();
-        if (tag == 0 || !mergeFieldFrom(tag, input)) {
-          break;
-        }
-      }
-      return this;
-    }
-
-    /**
-     * Parse a single field from {@code input} and merge it into this set.
-     * @param tag The field's tag number, which was already parsed.
-     * @return {@code false} if the tag is an end group tag.
-     */
-    public boolean mergeFieldFrom(final int tag, final CodedInputStream input)
-                                  throws IOException {
-      final int number = WireFormat.getTagFieldNumber(tag);
-      switch (WireFormat.getTagWireType(tag)) {
-        case WireFormat.WIRETYPE_VARINT:
-          getFieldBuilder(number).addVarint(input.readInt64());
-          return true;
-        case WireFormat.WIRETYPE_FIXED64:
-          getFieldBuilder(number).addFixed64(input.readFixed64());
-          return true;
-        case WireFormat.WIRETYPE_LENGTH_DELIMITED:
-          getFieldBuilder(number).addLengthDelimited(input.readBytes());
-          return true;
-        case WireFormat.WIRETYPE_START_GROUP:
-          final Builder subBuilder = newBuilder();
-          input.readGroup(number, subBuilder,
-                          ExtensionRegistry.getEmptyRegistry());
-          getFieldBuilder(number).addGroup(subBuilder.build());
-          return true;
-        case WireFormat.WIRETYPE_END_GROUP:
-          return false;
-        case WireFormat.WIRETYPE_FIXED32:
-          getFieldBuilder(number).addFixed32(input.readFixed32());
-          return true;
-        default:
-          throw InvalidProtocolBufferException.invalidWireType();
-      }
-    }
-
-    /**
-     * Parse {@code data} as an {@code UnknownFieldSet} and merge it with the
-     * set being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream)}.
-     */
-    @Override
-    public Builder mergeFrom(final ByteString data) throws InvalidProtocolBufferException {
-      try {
-        final CodedInputStream input = data.newCodedInput();
-        mergeFrom(input);
-        input.checkLastTagWas(0);
-        return this;
-      } catch (final InvalidProtocolBufferException e) {
-        throw e;
-      } catch (final IOException e) {
-        throw new RuntimeException(
-          "Reading from a ByteString threw an IOException (should " +
-          "never happen).", e);
-      }
-    }
-
-    /**
-     * Parse {@code data} as an {@code UnknownFieldSet} and merge it with the
-     * set being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream)}.
-     */
-    @Override
-    public Builder mergeFrom(final byte[] data) throws InvalidProtocolBufferException {
-      try {
-        final CodedInputStream input = CodedInputStream.newInstance(data);
-        mergeFrom(input);
-        input.checkLastTagWas(0);
-        return this;
-      } catch (final InvalidProtocolBufferException e) {
-        throw e;
-      } catch (final IOException e) {
-        throw new RuntimeException(
-          "Reading from a byte array threw an IOException (should " +
-          "never happen).", e);
-      }
-    }
-
-    /**
-     * Parse an {@code UnknownFieldSet} from {@code input} and merge it with the
-     * set being built.  This is just a small wrapper around
-     * {@link #mergeFrom(CodedInputStream)}.
-     */
-    @Override
-    public Builder mergeFrom(final InputStream input) throws IOException {
-      final CodedInputStream codedInput = CodedInputStream.newInstance(input);
-      mergeFrom(codedInput);
-      codedInput.checkLastTagWas(0);
-      return this;
-    }
-
-    @Override
-    public boolean mergeDelimitedFrom(InputStream input) throws IOException {
-      final int firstByte = input.read();
-      if (firstByte == -1) {
-        return false;
-      }
-      final int size = CodedInputStream.readRawVarint32(firstByte, input);
-      final InputStream limitedInput = new LimitedInputStream(input, size);
-      mergeFrom(limitedInput);
-      return true;
-    }
-
-    @Override
-    public boolean mergeDelimitedFrom(InputStream input, ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      // UnknownFieldSet has no extensions.
-      return mergeDelimitedFrom(input);
-    }
-
-    @Override
-    public Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      // UnknownFieldSet has no extensions.
-      return mergeFrom(input);
-    }
-
-    @Override
-    public Builder mergeFrom(ByteString data, ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      // UnknownFieldSet has no extensions.
-      return mergeFrom(data);
-    }
-
-    @Override
-    public Builder mergeFrom(byte[] data, int off, int len) throws InvalidProtocolBufferException {
-      try {
-        final CodedInputStream input =
-            CodedInputStream.newInstance(data, off, len);
-        mergeFrom(input);
-        input.checkLastTagWas(0);
-        return this;
-      } catch (InvalidProtocolBufferException e) {
-        throw e;
-      } catch (IOException e) {
-        throw new RuntimeException(
-          "Reading from a byte array threw an IOException (should " +
-          "never happen).", e);
-      }
-    }
-
-    @Override
-    public Builder mergeFrom(byte[] data, ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      // UnknownFieldSet has no extensions.
-      return mergeFrom(data);
-    }
-
-    @Override
-    public Builder mergeFrom(byte[] data, int off, int len, ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      // UnknownFieldSet has no extensions.
-      return mergeFrom(data, off, len);
-    }
-
-    @Override
-    public Builder mergeFrom(InputStream input, ExtensionRegistryLite extensionRegistry)
-        throws IOException {
-      // UnknownFieldSet has no extensions.
-      return mergeFrom(input);
-    }
-
-    @Override
-    public Builder mergeFrom(MessageLite m) {
-      if (m instanceof UnknownFieldSet) {
-        return mergeFrom((UnknownFieldSet) m);
-      }
-      throw new IllegalArgumentException(
-          "mergeFrom(MessageLite) can only merge messages of the same type.");
-    }
-
-    @Override
-    public boolean isInitialized() {
-      // UnknownFieldSets do not have required fields, so they are always
-      // initialized.
-      return true;
-    }
-  }
-
-  /**
-   * Represents a single field in an {@code UnknownFieldSet}.
-   *
-   * <p>A {@code Field} consists of five lists of values.  The lists correspond
-   * to the five "wire types" used in the protocol buffer binary format.
-   * The wire type of each field can be determined from the encoded form alone,
-   * without knowing the field's declared type.  So, we are able to parse
-   * unknown values at least this far and separate them.  Normally, only one
-   * of the five lists will contain any values, since it is impossible to
-   * define a valid message type that declares two different types for the
-   * same field number.  However, the code is designed to allow for the case
-   * where the same unknown field number is encountered using multiple different
-   * wire types.
-   *
-   * <p>{@code Field} is an immutable class.  To construct one, you must use a
-   * {@link Builder}.
-   *
-   * @see UnknownFieldSet
-   */
-  public static final class Field {
-    private Field() {}
-
-    /** Construct a new {@link Builder}. */
-    public static Builder newBuilder() {
-      return Builder.create();
-    }
-
-    /**
-     * Construct a new {@link Builder} and initialize it to a copy of
-     * {@code copyFrom}.
-     */
-    public static Builder newBuilder(final Field copyFrom) {
-      return newBuilder().mergeFrom(copyFrom);
-    }
-
-    /** Get an empty {@code Field}. */
-    public static Field getDefaultInstance() {
-      return fieldDefaultInstance;
-    }
-    private static final Field fieldDefaultInstance = newBuilder().build();
-
-    /** Get the list of varint values for this field. */
-    public List<Long> getVarintList()               { return varint;          }
-
-    /** Get the list of fixed32 values for this field. */
-    public List<Integer> getFixed32List()           { return fixed32;         }
-
-    /** Get the list of fixed64 values for this field. */
-    public List<Long> getFixed64List()              { return fixed64;         }
-
-    /** Get the list of length-delimited values for this field. */
-    public List<ByteString> getLengthDelimitedList() { return lengthDelimited; }
-
-    /**
-     * Get the list of embedded group values for this field.  These are
-     * represented using {@link UnknownFieldSet}s rather than {@link Message}s
-     * since the group's type is presumably unknown.
-     */
-    public List<UnknownFieldSet> getGroupList()      { return group;           }
-
-    @Override
-    public boolean equals(final Object other) {
-      if (this == other) {
-        return true;
-      }
-      if (!(other instanceof Field)) {
-        return false;
-      }
-      return Arrays.equals(getIdentityArray(),
-          ((Field) other).getIdentityArray());
-    }
-
-    @Override
-    public int hashCode() {
-      return Arrays.hashCode(getIdentityArray());
-    }
-
-    /**
-     * Returns the array of objects to be used to uniquely identify this
-     * {@link Field} instance.
-     */
-    private Object[] getIdentityArray() {
-      return new Object[] {
-          varint,
-          fixed32,
-          fixed64,
-          lengthDelimited,
-          group};
-    }
-
-    /**
-     * Serializes the field, including field number, and writes it to
-     * {@code output}.
-     */
-    public void writeTo(final int fieldNumber, final CodedOutputStream output)
-                        throws IOException {
-      for (final long value : varint) {
-        output.writeUInt64(fieldNumber, value);
-      }
-      for (final int value : fixed32) {
-        output.writeFixed32(fieldNumber, value);
-      }
-      for (final long value : fixed64) {
-        output.writeFixed64(fieldNumber, value);
-      }
-      for (final ByteString value : lengthDelimited) {
-        output.writeBytes(fieldNumber, value);
-      }
-      for (final UnknownFieldSet value : group) {
-        output.writeGroup(fieldNumber, value);
-      }
-    }
-
-    /**
-     * Get the number of bytes required to encode this field, including field
-     * number.
-     */
-    public int getSerializedSize(final int fieldNumber) {
-      int result = 0;
-      for (final long value : varint) {
-        result += CodedOutputStream.computeUInt64Size(fieldNumber, value);
-      }
-      for (final int value : fixed32) {
-        result += CodedOutputStream.computeFixed32Size(fieldNumber, value);
-      }
-      for (final long value : fixed64) {
-        result += CodedOutputStream.computeFixed64Size(fieldNumber, value);
-      }
-      for (final ByteString value : lengthDelimited) {
-        result += CodedOutputStream.computeBytesSize(fieldNumber, value);
-      }
-      for (final UnknownFieldSet value : group) {
-        result += CodedOutputStream.computeGroupSize(fieldNumber, value);
-      }
-      return result;
-    }
-
-    /**
-     * Serializes the field, including field number, and writes it to
-     * {@code output}, using {@code MessageSet} wire format.
-     */
-    public void writeAsMessageSetExtensionTo(
-        final int fieldNumber,
-        final CodedOutputStream output)
-        throws IOException {
-      for (final ByteString value : lengthDelimited) {
-        output.writeRawMessageSetExtension(fieldNumber, value);
-      }
-    }
-
-
-    /**
-     * Get the number of bytes required to encode this field, including field number, using {@code
-     * MessageSet} wire format.
-     */
-    public int getSerializedSizeAsMessageSetExtension(final int fieldNumber) {
-      int result = 0;
-      for (final ByteString value : lengthDelimited) {
-        result += CodedOutputStream.computeRawMessageSetExtensionSize(
-          fieldNumber, value);
-      }
-      return result;
-    }
-
-    private List<Long> varint;
-    private List<Integer> fixed32;
-    private List<Long> fixed64;
-    private List<ByteString> lengthDelimited;
-    private List<UnknownFieldSet> group;
-
-    /**
-     * Used to build a {@link Field} within an {@link UnknownFieldSet}.
-     *
-     * <p>Use {@link Field#newBuilder()} to construct a {@code Builder}.
-     */
-    public static final class Builder {
-      // This constructor should never be called directly (except from 'create').
-      private Builder() {}
-
-      private static Builder create() {
-        Builder builder = new Builder();
-        builder.result = new Field();
-        return builder;
-      }
-
-      private Field result;
-
-      /**
-       * Build the field.  After {@code build()} has been called, the
-       * {@code Builder} is no longer usable.  Calling any other method will
-       * result in undefined behavior and can cause a
-       * {@code NullPointerException} to be thrown.
-       */
-      public Field build() {
-        if (result.varint == null) {
-          result.varint = Collections.emptyList();
-        } else {
-          result.varint = Collections.unmodifiableList(result.varint);
-        }
-        if (result.fixed32 == null) {
-          result.fixed32 = Collections.emptyList();
-        } else {
-          result.fixed32 = Collections.unmodifiableList(result.fixed32);
-        }
-        if (result.fixed64 == null) {
-          result.fixed64 = Collections.emptyList();
-        } else {
-          result.fixed64 = Collections.unmodifiableList(result.fixed64);
-        }
-        if (result.lengthDelimited == null) {
-          result.lengthDelimited = Collections.emptyList();
-        } else {
-          result.lengthDelimited =
-            Collections.unmodifiableList(result.lengthDelimited);
-        }
-        if (result.group == null) {
-          result.group = Collections.emptyList();
-        } else {
-          result.group = Collections.unmodifiableList(result.group);
-        }
-
-        final Field returnMe = result;
-        result = null;
-        return returnMe;
-      }
-
-      /** Discard the field's contents. */
-      public Builder clear() {
-        result = new Field();
-        return this;
-      }
-
-      /**
-       * Merge the values in {@code other} into this field.  For each list
-       * of values, {@code other}'s values are append to the ones in this
-       * field.
-       */
-      public Builder mergeFrom(final Field other) {
-        if (!other.varint.isEmpty()) {
-          if (result.varint == null) {
-            result.varint = new ArrayList<Long>();
-          }
-          result.varint.addAll(other.varint);
-        }
-        if (!other.fixed32.isEmpty()) {
-          if (result.fixed32 == null) {
-            result.fixed32 = new ArrayList<Integer>();
-          }
-          result.fixed32.addAll(other.fixed32);
-        }
-        if (!other.fixed64.isEmpty()) {
-          if (result.fixed64 == null) {
-            result.fixed64 = new ArrayList<Long>();
-          }
-          result.fixed64.addAll(other.fixed64);
-        }
-        if (!other.lengthDelimited.isEmpty()) {
-          if (result.lengthDelimited == null) {
-            result.lengthDelimited = new ArrayList<ByteString>();
-          }
-          result.lengthDelimited.addAll(other.lengthDelimited);
-        }
-        if (!other.group.isEmpty()) {
-          if (result.group == null) {
-            result.group = new ArrayList<UnknownFieldSet>();
-          }
-          result.group.addAll(other.group);
-        }
-        return this;
-      }
-
-      /** Add a varint value. */
-      public Builder addVarint(final long value) {
-        if (result.varint == null) {
-          result.varint = new ArrayList<Long>();
-        }
-        result.varint.add(value);
-        return this;
-      }
-
-      /** Add a fixed32 value. */
-      public Builder addFixed32(final int value) {
-        if (result.fixed32 == null) {
-          result.fixed32 = new ArrayList<Integer>();
-        }
-        result.fixed32.add(value);
-        return this;
-      }
-
-      /** Add a fixed64 value. */
-      public Builder addFixed64(final long value) {
-        if (result.fixed64 == null) {
-          result.fixed64 = new ArrayList<Long>();
-        }
-        result.fixed64.add(value);
-        return this;
-      }
-
-      /** Add a length-delimited value. */
-      public Builder addLengthDelimited(final ByteString value) {
-        if (result.lengthDelimited == null) {
-          result.lengthDelimited = new ArrayList<ByteString>();
-        }
-        result.lengthDelimited.add(value);
-        return this;
-      }
-
-      /** Add an embedded group. */
-      public Builder addGroup(final UnknownFieldSet value) {
-        if (result.group == null) {
-          result.group = new ArrayList<UnknownFieldSet>();
-        }
-        result.group.add(value);
-        return this;
-      }
-    }
-  }
-
-  /**
-   * Parser to implement MessageLite interface.
-   */
-  public static final class Parser extends AbstractParser<UnknownFieldSet> {
-    @Override
-    public UnknownFieldSet parsePartialFrom(
-        CodedInputStream input, ExtensionRegistryLite extensionRegistry)
-        throws InvalidProtocolBufferException {
-      Builder builder = newBuilder();
-      try {
-        builder.mergeFrom(input);
-      } catch (InvalidProtocolBufferException e) {
-        throw e.setUnfinishedMessage(builder.buildPartial());
-      } catch (IOException e) {
-        throw new InvalidProtocolBufferException(e)
-            .setUnfinishedMessage(builder.buildPartial());
-      }
-      return builder.buildPartial();
-    }
-  }
-
-  private static final Parser PARSER = new Parser();
-  @Override
-  public final Parser getParserForType() {
-    return PARSER;
-  }
-}


[40/52] [partial] hbase git commit: HBASE-17056 Remove checked in PB generated files

Posted by st...@apache.org.
http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Any.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Any.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Any.java
deleted file mode 100644
index bf8e852..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/Any.java
+++ /dev/null
@@ -1,899 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/any.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-/**
- * <pre>
- * `Any` contains an arbitrary serialized protocol buffer message along with a
- * URL that describes the type of the serialized message.
- * Protobuf library provides support to pack/unpack Any values in the form
- * of utility functions or additional generated methods of the Any type.
- * Example 1: Pack and unpack a message in C++.
- *     Foo foo = ...;
- *     Any any;
- *     any.PackFrom(foo);
- *     ...
- *     if (any.UnpackTo(&amp;foo)) {
- *       ...
- *     }
- * Example 2: Pack and unpack a message in Java.
- *     Foo foo = ...;
- *     Any any = Any.pack(foo);
- *     ...
- *     if (any.is(Foo.class)) {
- *       foo = any.unpack(Foo.class);
- *     }
- *  Example 3: Pack and unpack a message in Python.
- *     foo = Foo(...)
- *     any = Any()
- *     any.Pack(foo)
- *     ...
- *     if any.Is(Foo.DESCRIPTOR):
- *       any.Unpack(foo)
- *       ...
- * The pack methods provided by protobuf library will by default use
- * 'type.googleapis.com/full.type.name' as the type URL and the unpack
- * methods only use the fully qualified type name after the last '/'
- * in the type URL, for example "foo.bar.com/x/y.z" will yield type
- * name "y.z".
- * JSON
- * ====
- * The JSON representation of an `Any` value uses the regular
- * representation of the deserialized, embedded message, with an
- * additional field `&#64;type` which contains the type URL. Example:
- *     package google.profile;
- *     message Person {
- *       string first_name = 1;
- *       string last_name = 2;
- *     }
- *     {
- *       "&#64;type": "type.googleapis.com/google.profile.Person",
- *       "firstName": &lt;string&gt;,
- *       "lastName": &lt;string&gt;
- *     }
- * If the embedded message type is well-known and has a custom JSON
- * representation, that representation will be embedded adding a field
- * `value` which holds the custom JSON in addition to the `&#64;type`
- * field. Example (for message [google.protobuf.Duration][]):
- *     {
- *       "&#64;type": "type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration",
- *       "value": "1.212s"
- *     }
- * </pre>
- *
- * Protobuf type {@code google.protobuf.Any}
- */
-public  final class Any extends
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements
-    // @@protoc_insertion_point(message_implements:google.protobuf.Any)
-    AnyOrBuilder {
-  // Use Any.newBuilder() to construct.
-  private Any(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
-    super(builder);
-  }
-  private Any() {
-    typeUrl_ = "";
-    value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-  }
-
-  @java.lang.Override
-  public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet
-  getUnknownFields() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance();
-  }
-  private Any(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    this();
-    int mutable_bitField0_ = 0;
-    try {
-      boolean done = false;
-      while (!done) {
-        int tag = input.readTag();
-        switch (tag) {
-          case 0:
-            done = true;
-            break;
-          default: {
-            if (!input.skipField(tag)) {
-              done = true;
-            }
-            break;
-          }
-          case 10: {
-            java.lang.String s = input.readStringRequireUtf8();
-
-            typeUrl_ = s;
-            break;
-          }
-          case 18: {
-
-            value_ = input.readBytes();
-            break;
-          }
-        }
-      }
-    } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-      throw e.setUnfinishedMessage(this);
-    } catch (java.io.IOException e) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          e).setUnfinishedMessage(this);
-    } finally {
-      makeExtensionsImmutable();
-    }
-  }
-  public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-      getDescriptor() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyProto.internal_static_google_protobuf_Any_descriptor;
-  }
-
-  protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internalGetFieldAccessorTable() {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyProto.internal_static_google_protobuf_Any_fieldAccessorTable
-        .ensureFieldAccessorsInitialized(
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.Builder.class);
-  }
-
-  private static String getTypeUrl(
-      java.lang.String typeUrlPrefix,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor descriptor) {
-    return typeUrlPrefix.endsWith("/")
-        ? typeUrlPrefix + descriptor.getFullName()
-        : typeUrlPrefix + "/" + descriptor.getFullName();
-  }
-
-  private static String getTypeNameFromTypeUrl(
-      java.lang.String typeUrl) {
-    int pos = typeUrl.lastIndexOf('/');
-    return pos == -1 ? "" : typeUrl.substring(pos + 1);
-  }
-
-  public static <T extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Message> Any pack(
-      T message) {
-    return Any.newBuilder()
-        .setTypeUrl(getTypeUrl("type.googleapis.com",
-                               message.getDescriptorForType()))
-        .setValue(message.toByteString())
-        .build();
-  }
-
-  /**
-   * Packs a message using the given type URL prefix. The type URL will
-   * be constructed by concatenating the message type's full name to the
-   * prefix with an optional "/" separator if the prefix doesn't end
-   * with "/" already.
-   */
-  public static <T extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Message> Any pack(
-      T message, java.lang.String typeUrlPrefix) {
-    return Any.newBuilder()
-        .setTypeUrl(getTypeUrl(typeUrlPrefix,
-                               message.getDescriptorForType()))
-        .setValue(message.toByteString())
-        .build();
-  }
-
-  public <T extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Message> boolean is(
-      java.lang.Class<T> clazz) {
-    T defaultInstance =
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.getDefaultInstance(clazz);
-    return getTypeNameFromTypeUrl(getTypeUrl()).equals(
-        defaultInstance.getDescriptorForType().getFullName());
-  }
-
-  private volatile org.apache.hadoop.hbase.shaded.com.google.protobuf.Message cachedUnpackValue;
-
-  public <T extends org.apache.hadoop.hbase.shaded.com.google.protobuf.Message> T unpack(
-      java.lang.Class<T> clazz)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    if (!is(clazz)) {
-      throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException(
-          "Type of the Any message does not match the given class.");
-    }
-    if (cachedUnpackValue != null) {
-      return (T) cachedUnpackValue;
-    }
-    T defaultInstance =
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.getDefaultInstance(clazz);
-    T result = (T) defaultInstance.getParserForType()
-        .parseFrom(getValue());
-    cachedUnpackValue = result;
-    return result;
-  }
-  public static final int TYPE_URL_FIELD_NUMBER = 1;
-  private volatile java.lang.Object typeUrl_;
-  /**
-   * <pre>
-   * A URL/resource name whose content describes the type of the
-   * serialized protocol buffer message.
-   * For URLs which use the scheme `http`, `https`, or no scheme, the
-   * following restrictions and interpretations apply:
-   * * If no scheme is provided, `https` is assumed.
-   * * The last segment of the URL's path must represent the fully
-   *   qualified name of the type (as in `path/google.protobuf.Duration`).
-   *   The name should be in a canonical form (e.g., leading "." is
-   *   not accepted).
-   * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-   *   value in binary format, or produce an error.
-   * * Applications are allowed to cache lookup results based on the
-   *   URL, or have them precompiled into a binary to avoid any
-   *   lookup. Therefore, binary compatibility needs to be preserved
-   *   on changes to types. (Use versioned type names to manage
-   *   breaking changes.)
-   * Schemes other than `http`, `https` (or the empty scheme) might be
-   * used with implementation specific semantics.
-   * </pre>
-   *
-   * <code>string type_url = 1;</code>
-   */
-  public java.lang.String getTypeUrl() {
-    java.lang.Object ref = typeUrl_;
-    if (ref instanceof java.lang.String) {
-      return (java.lang.String) ref;
-    } else {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs = 
-          (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      java.lang.String s = bs.toStringUtf8();
-      typeUrl_ = s;
-      return s;
-    }
-  }
-  /**
-   * <pre>
-   * A URL/resource name whose content describes the type of the
-   * serialized protocol buffer message.
-   * For URLs which use the scheme `http`, `https`, or no scheme, the
-   * following restrictions and interpretations apply:
-   * * If no scheme is provided, `https` is assumed.
-   * * The last segment of the URL's path must represent the fully
-   *   qualified name of the type (as in `path/google.protobuf.Duration`).
-   *   The name should be in a canonical form (e.g., leading "." is
-   *   not accepted).
-   * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-   *   value in binary format, or produce an error.
-   * * Applications are allowed to cache lookup results based on the
-   *   URL, or have them precompiled into a binary to avoid any
-   *   lookup. Therefore, binary compatibility needs to be preserved
-   *   on changes to types. (Use versioned type names to manage
-   *   breaking changes.)
-   * Schemes other than `http`, `https` (or the empty scheme) might be
-   * used with implementation specific semantics.
-   * </pre>
-   *
-   * <code>string type_url = 1;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getTypeUrlBytes() {
-    java.lang.Object ref = typeUrl_;
-    if (ref instanceof java.lang.String) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-          org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-              (java.lang.String) ref);
-      typeUrl_ = b;
-      return b;
-    } else {
-      return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-    }
-  }
-
-  public static final int VALUE_FIELD_NUMBER = 2;
-  private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_;
-  /**
-   * <pre>
-   * Must be a valid serialized protocol buffer of the above specified type.
-   * </pre>
-   *
-   * <code>bytes value = 2;</code>
-   */
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
-    return value_;
-  }
-
-  private byte memoizedIsInitialized = -1;
-  public final boolean isInitialized() {
-    byte isInitialized = memoizedIsInitialized;
-    if (isInitialized == 1) return true;
-    if (isInitialized == 0) return false;
-
-    memoizedIsInitialized = 1;
-    return true;
-  }
-
-  public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output)
-                      throws java.io.IOException {
-    if (!getTypeUrlBytes().isEmpty()) {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, typeUrl_);
-    }
-    if (!value_.isEmpty()) {
-      output.writeBytes(2, value_);
-    }
-  }
-
-  public int getSerializedSize() {
-    int size = memoizedSize;
-    if (size != -1) return size;
-
-    size = 0;
-    if (!getTypeUrlBytes().isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, typeUrl_);
-    }
-    if (!value_.isEmpty()) {
-      size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream
-        .computeBytesSize(2, value_);
-    }
-    memoizedSize = size;
-    return size;
-  }
-
-  private static final long serialVersionUID = 0L;
-  @java.lang.Override
-  public boolean equals(final java.lang.Object obj) {
-    if (obj == this) {
-     return true;
-    }
-    if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Any)) {
-      return super.equals(obj);
-    }
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Any other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Any) obj;
-
-    boolean result = true;
-    result = result && getTypeUrl()
-        .equals(other.getTypeUrl());
-    result = result && getValue()
-        .equals(other.getValue());
-    return result;
-  }
-
-  @java.lang.Override
-  public int hashCode() {
-    if (memoizedHashCode != 0) {
-      return memoizedHashCode;
-    }
-    int hash = 41;
-    hash = (19 * hash) + getDescriptor().hashCode();
-    hash = (37 * hash) + TYPE_URL_FIELD_NUMBER;
-    hash = (53 * hash) + getTypeUrl().hashCode();
-    hash = (37 * hash) + VALUE_FIELD_NUMBER;
-    hash = (53 * hash) + getValue().hashCode();
-    hash = (29 * hash) + unknownFields.hashCode();
-    memoizedHashCode = hash;
-    return hash;
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseFrom(byte[] data)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseFrom(
-      byte[] data,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-    return PARSER.parseFrom(data, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseDelimitedFrom(java.io.InputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseDelimitedFrom(
-      java.io.InputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input);
-  }
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parseFrom(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-      throws java.io.IOException {
-    return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-        .parseWithIOException(PARSER, input, extensionRegistry);
-  }
-
-  public Builder newBuilderForType() { return newBuilder(); }
-  public static Builder newBuilder() {
-    return DEFAULT_INSTANCE.toBuilder();
-  }
-  public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.Any prototype) {
-    return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
-  }
-  public Builder toBuilder() {
-    return this == DEFAULT_INSTANCE
-        ? new Builder() : new Builder().mergeFrom(this);
-  }
-
-  @java.lang.Override
-  protected Builder newBuilderForType(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-    Builder builder = new Builder(parent);
-    return builder;
-  }
-  /**
-   * <pre>
-   * `Any` contains an arbitrary serialized protocol buffer message along with a
-   * URL that describes the type of the serialized message.
-   * Protobuf library provides support to pack/unpack Any values in the form
-   * of utility functions or additional generated methods of the Any type.
-   * Example 1: Pack and unpack a message in C++.
-   *     Foo foo = ...;
-   *     Any any;
-   *     any.PackFrom(foo);
-   *     ...
-   *     if (any.UnpackTo(&amp;foo)) {
-   *       ...
-   *     }
-   * Example 2: Pack and unpack a message in Java.
-   *     Foo foo = ...;
-   *     Any any = Any.pack(foo);
-   *     ...
-   *     if (any.is(Foo.class)) {
-   *       foo = any.unpack(Foo.class);
-   *     }
-   *  Example 3: Pack and unpack a message in Python.
-   *     foo = Foo(...)
-   *     any = Any()
-   *     any.Pack(foo)
-   *     ...
-   *     if any.Is(Foo.DESCRIPTOR):
-   *       any.Unpack(foo)
-   *       ...
-   * The pack methods provided by protobuf library will by default use
-   * 'type.googleapis.com/full.type.name' as the type URL and the unpack
-   * methods only use the fully qualified type name after the last '/'
-   * in the type URL, for example "foo.bar.com/x/y.z" will yield type
-   * name "y.z".
-   * JSON
-   * ====
-   * The JSON representation of an `Any` value uses the regular
-   * representation of the deserialized, embedded message, with an
-   * additional field `&#64;type` which contains the type URL. Example:
-   *     package google.profile;
-   *     message Person {
-   *       string first_name = 1;
-   *       string last_name = 2;
-   *     }
-   *     {
-   *       "&#64;type": "type.googleapis.com/google.profile.Person",
-   *       "firstName": &lt;string&gt;,
-   *       "lastName": &lt;string&gt;
-   *     }
-   * If the embedded message type is well-known and has a custom JSON
-   * representation, that representation will be embedded adding a field
-   * `value` which holds the custom JSON in addition to the `&#64;type`
-   * field. Example (for message [google.protobuf.Duration][]):
-   *     {
-   *       "&#64;type": "type.googleapis.org.apache.hadoop.hbase.shaded.com.google.protobuf.Duration",
-   *       "value": "1.212s"
-   *     }
-   * </pre>
-   *
-   * Protobuf type {@code google.protobuf.Any}
-   */
-  public static final class Builder extends
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
-      // @@protoc_insertion_point(builder_implements:google.protobuf.Any)
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyOrBuilder {
-    public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyProto.internal_static_google_protobuf_Any_descriptor;
-    }
-
-    protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-        internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyProto.internal_static_google_protobuf_Any_fieldAccessorTable
-          .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.Builder.class);
-    }
-
-    // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.newBuilder()
-    private Builder() {
-      maybeForceBuilderInitialization();
-    }
-
-    private Builder(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
-      super(parent);
-      maybeForceBuilderInitialization();
-    }
-    private void maybeForceBuilderInitialization() {
-      if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3
-              .alwaysUseFieldBuilders) {
-      }
-    }
-    public Builder clear() {
-      super.clear();
-      typeUrl_ = "";
-
-      value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-
-      return this;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-        getDescriptorForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.AnyProto.internal_static_google_protobuf_Any_descriptor;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Any getDefaultInstanceForType() {
-      return org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.getDefaultInstance();
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Any build() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Any result = buildPartial();
-      if (!result.isInitialized()) {
-        throw newUninitializedMessageException(result);
-      }
-      return result;
-    }
-
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.Any buildPartial() {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Any result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Any(this);
-      result.typeUrl_ = typeUrl_;
-      result.value_ = value_;
-      onBuilt();
-      return result;
-    }
-
-    public Builder clone() {
-      return (Builder) super.clone();
-    }
-    public Builder setField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.setField(field, value);
-    }
-    public Builder clearField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) {
-      return (Builder) super.clearField(field);
-    }
-    public Builder clearOneof(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
-      return (Builder) super.clearOneof(oneof);
-    }
-    public Builder setRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        int index, Object value) {
-      return (Builder) super.setRepeatedField(field, index, value);
-    }
-    public Builder addRepeatedField(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field,
-        Object value) {
-      return (Builder) super.addRepeatedField(field, value);
-    }
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) {
-      if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.Any) {
-        return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.Any)other);
-      } else {
-        super.mergeFrom(other);
-        return this;
-      }
-    }
-
-    public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Any other) {
-      if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.Any.getDefaultInstance()) return this;
-      if (!other.getTypeUrl().isEmpty()) {
-        typeUrl_ = other.typeUrl_;
-        onChanged();
-      }
-      if (other.getValue() != org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY) {
-        setValue(other.getValue());
-      }
-      onChanged();
-      return this;
-    }
-
-    public final boolean isInitialized() {
-      return true;
-    }
-
-    public Builder mergeFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws java.io.IOException {
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.Any parsedMessage = null;
-      try {
-        parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
-      } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) {
-        parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.Any) e.getUnfinishedMessage();
-        throw e.unwrapIOException();
-      } finally {
-        if (parsedMessage != null) {
-          mergeFrom(parsedMessage);
-        }
-      }
-      return this;
-    }
-
-    private java.lang.Object typeUrl_ = "";
-    /**
-     * <pre>
-     * A URL/resource name whose content describes the type of the
-     * serialized protocol buffer message.
-     * For URLs which use the scheme `http`, `https`, or no scheme, the
-     * following restrictions and interpretations apply:
-     * * If no scheme is provided, `https` is assumed.
-     * * The last segment of the URL's path must represent the fully
-     *   qualified name of the type (as in `path/google.protobuf.Duration`).
-     *   The name should be in a canonical form (e.g., leading "." is
-     *   not accepted).
-     * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-     *   value in binary format, or produce an error.
-     * * Applications are allowed to cache lookup results based on the
-     *   URL, or have them precompiled into a binary to avoid any
-     *   lookup. Therefore, binary compatibility needs to be preserved
-     *   on changes to types. (Use versioned type names to manage
-     *   breaking changes.)
-     * Schemes other than `http`, `https` (or the empty scheme) might be
-     * used with implementation specific semantics.
-     * </pre>
-     *
-     * <code>string type_url = 1;</code>
-     */
-    public java.lang.String getTypeUrl() {
-      java.lang.Object ref = typeUrl_;
-      if (!(ref instanceof java.lang.String)) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString bs =
-            (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-        java.lang.String s = bs.toStringUtf8();
-        typeUrl_ = s;
-        return s;
-      } else {
-        return (java.lang.String) ref;
-      }
-    }
-    /**
-     * <pre>
-     * A URL/resource name whose content describes the type of the
-     * serialized protocol buffer message.
-     * For URLs which use the scheme `http`, `https`, or no scheme, the
-     * following restrictions and interpretations apply:
-     * * If no scheme is provided, `https` is assumed.
-     * * The last segment of the URL's path must represent the fully
-     *   qualified name of the type (as in `path/google.protobuf.Duration`).
-     *   The name should be in a canonical form (e.g., leading "." is
-     *   not accepted).
-     * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-     *   value in binary format, or produce an error.
-     * * Applications are allowed to cache lookup results based on the
-     *   URL, or have them precompiled into a binary to avoid any
-     *   lookup. Therefore, binary compatibility needs to be preserved
-     *   on changes to types. (Use versioned type names to manage
-     *   breaking changes.)
-     * Schemes other than `http`, `https` (or the empty scheme) might be
-     * used with implementation specific semantics.
-     * </pre>
-     *
-     * <code>string type_url = 1;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-        getTypeUrlBytes() {
-      java.lang.Object ref = typeUrl_;
-      if (ref instanceof String) {
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString b = 
-            org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.copyFromUtf8(
-                (java.lang.String) ref);
-        typeUrl_ = b;
-        return b;
-      } else {
-        return (org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString) ref;
-      }
-    }
-    /**
-     * <pre>
-     * A URL/resource name whose content describes the type of the
-     * serialized protocol buffer message.
-     * For URLs which use the scheme `http`, `https`, or no scheme, the
-     * following restrictions and interpretations apply:
-     * * If no scheme is provided, `https` is assumed.
-     * * The last segment of the URL's path must represent the fully
-     *   qualified name of the type (as in `path/google.protobuf.Duration`).
-     *   The name should be in a canonical form (e.g., leading "." is
-     *   not accepted).
-     * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-     *   value in binary format, or produce an error.
-     * * Applications are allowed to cache lookup results based on the
-     *   URL, or have them precompiled into a binary to avoid any
-     *   lookup. Therefore, binary compatibility needs to be preserved
-     *   on changes to types. (Use versioned type names to manage
-     *   breaking changes.)
-     * Schemes other than `http`, `https` (or the empty scheme) might be
-     * used with implementation specific semantics.
-     * </pre>
-     *
-     * <code>string type_url = 1;</code>
-     */
-    public Builder setTypeUrl(
-        java.lang.String value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      typeUrl_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * A URL/resource name whose content describes the type of the
-     * serialized protocol buffer message.
-     * For URLs which use the scheme `http`, `https`, or no scheme, the
-     * following restrictions and interpretations apply:
-     * * If no scheme is provided, `https` is assumed.
-     * * The last segment of the URL's path must represent the fully
-     *   qualified name of the type (as in `path/google.protobuf.Duration`).
-     *   The name should be in a canonical form (e.g., leading "." is
-     *   not accepted).
-     * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-     *   value in binary format, or produce an error.
-     * * Applications are allowed to cache lookup results based on the
-     *   URL, or have them precompiled into a binary to avoid any
-     *   lookup. Therefore, binary compatibility needs to be preserved
-     *   on changes to types. (Use versioned type names to manage
-     *   breaking changes.)
-     * Schemes other than `http`, `https` (or the empty scheme) might be
-     * used with implementation specific semantics.
-     * </pre>
-     *
-     * <code>string type_url = 1;</code>
-     */
-    public Builder clearTypeUrl() {
-      
-      typeUrl_ = getDefaultInstance().getTypeUrl();
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * A URL/resource name whose content describes the type of the
-     * serialized protocol buffer message.
-     * For URLs which use the scheme `http`, `https`, or no scheme, the
-     * following restrictions and interpretations apply:
-     * * If no scheme is provided, `https` is assumed.
-     * * The last segment of the URL's path must represent the fully
-     *   qualified name of the type (as in `path/google.protobuf.Duration`).
-     *   The name should be in a canonical form (e.g., leading "." is
-     *   not accepted).
-     * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-     *   value in binary format, or produce an error.
-     * * Applications are allowed to cache lookup results based on the
-     *   URL, or have them precompiled into a binary to avoid any
-     *   lookup. Therefore, binary compatibility needs to be preserved
-     *   on changes to types. (Use versioned type names to manage
-     *   breaking changes.)
-     * Schemes other than `http`, `https` (or the empty scheme) might be
-     * used with implementation specific semantics.
-     * </pre>
-     *
-     * <code>string type_url = 1;</code>
-     */
-    public Builder setTypeUrlBytes(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  checkByteStringIsUtf8(value);
-      
-      typeUrl_ = value;
-      onChanged();
-      return this;
-    }
-
-    private org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value_ = org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString.EMPTY;
-    /**
-     * <pre>
-     * Must be a valid serialized protocol buffer of the above specified type.
-     * </pre>
-     *
-     * <code>bytes value = 2;</code>
-     */
-    public org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue() {
-      return value_;
-    }
-    /**
-     * <pre>
-     * Must be a valid serialized protocol buffer of the above specified type.
-     * </pre>
-     *
-     * <code>bytes value = 2;</code>
-     */
-    public Builder setValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString value) {
-      if (value == null) {
-    throw new NullPointerException();
-  }
-  
-      value_ = value;
-      onChanged();
-      return this;
-    }
-    /**
-     * <pre>
-     * Must be a valid serialized protocol buffer of the above specified type.
-     * </pre>
-     *
-     * <code>bytes value = 2;</code>
-     */
-    public Builder clearValue() {
-      
-      value_ = getDefaultInstance().getValue();
-      onChanged();
-      return this;
-    }
-    public final Builder setUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-    public final Builder mergeUnknownFields(
-        final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) {
-      return this;
-    }
-
-
-    // @@protoc_insertion_point(builder_scope:google.protobuf.Any)
-  }
-
-  // @@protoc_insertion_point(class_scope:google.protobuf.Any)
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Any DEFAULT_INSTANCE;
-  static {
-    DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.Any();
-  }
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Any getDefaultInstance() {
-    return DEFAULT_INSTANCE;
-  }
-
-  private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Any>
-      PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<Any>() {
-    public Any parsePartialFrom(
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input,
-        org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-        throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException {
-        return new Any(input, extensionRegistry);
-    }
-  };
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Any> parser() {
-    return PARSER;
-  }
-
-  @java.lang.Override
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<Any> getParserForType() {
-    return PARSER;
-  }
-
-  public org.apache.hadoop.hbase.shaded.com.google.protobuf.Any getDefaultInstanceForType() {
-    return DEFAULT_INSTANCE;
-  }
-
-}
-

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AnyOrBuilder.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AnyOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AnyOrBuilder.java
deleted file mode 100644
index 80981c8..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AnyOrBuilder.java
+++ /dev/null
@@ -1,70 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/any.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public interface AnyOrBuilder extends
-    // @@protoc_insertion_point(interface_extends:google.protobuf.Any)
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder {
-
-  /**
-   * <pre>
-   * A URL/resource name whose content describes the type of the
-   * serialized protocol buffer message.
-   * For URLs which use the scheme `http`, `https`, or no scheme, the
-   * following restrictions and interpretations apply:
-   * * If no scheme is provided, `https` is assumed.
-   * * The last segment of the URL's path must represent the fully
-   *   qualified name of the type (as in `path/google.protobuf.Duration`).
-   *   The name should be in a canonical form (e.g., leading "." is
-   *   not accepted).
-   * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-   *   value in binary format, or produce an error.
-   * * Applications are allowed to cache lookup results based on the
-   *   URL, or have them precompiled into a binary to avoid any
-   *   lookup. Therefore, binary compatibility needs to be preserved
-   *   on changes to types. (Use versioned type names to manage
-   *   breaking changes.)
-   * Schemes other than `http`, `https` (or the empty scheme) might be
-   * used with implementation specific semantics.
-   * </pre>
-   *
-   * <code>string type_url = 1;</code>
-   */
-  java.lang.String getTypeUrl();
-  /**
-   * <pre>
-   * A URL/resource name whose content describes the type of the
-   * serialized protocol buffer message.
-   * For URLs which use the scheme `http`, `https`, or no scheme, the
-   * following restrictions and interpretations apply:
-   * * If no scheme is provided, `https` is assumed.
-   * * The last segment of the URL's path must represent the fully
-   *   qualified name of the type (as in `path/google.protobuf.Duration`).
-   *   The name should be in a canonical form (e.g., leading "." is
-   *   not accepted).
-   * * An HTTP GET on the URL must yield a [google.protobuf.Type][]
-   *   value in binary format, or produce an error.
-   * * Applications are allowed to cache lookup results based on the
-   *   URL, or have them precompiled into a binary to avoid any
-   *   lookup. Therefore, binary compatibility needs to be preserved
-   *   on changes to types. (Use versioned type names to manage
-   *   breaking changes.)
-   * Schemes other than `http`, `https` (or the empty scheme) might be
-   * used with implementation specific semantics.
-   * </pre>
-   *
-   * <code>string type_url = 1;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString
-      getTypeUrlBytes();
-
-  /**
-   * <pre>
-   * Must be a valid serialized protocol buffer of the above specified type.
-   * </pre>
-   *
-   * <code>bytes value = 2;</code>
-   */
-  org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString getValue();
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/ee70b1d2/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AnyProto.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AnyProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AnyProto.java
deleted file mode 100644
index cc674c5..0000000
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/AnyProto.java
+++ /dev/null
@@ -1,59 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: google/protobuf/any.proto
-
-package org.apache.hadoop.hbase.shaded.com.google.protobuf;
-
-public final class AnyProto {
-  private AnyProto() {}
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) {
-  }
-
-  public static void registerAllExtensions(
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) {
-    registerAllExtensions(
-        (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry);
-  }
-  static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor
-    internal_static_google_protobuf_Any_descriptor;
-  static final 
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
-      internal_static_google_protobuf_Any_fieldAccessorTable;
-
-  public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static  org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\031google/protobuf/any.proto\022\017google.prot" +
-      "obuf\"&\n\003Any\022\020\n\010type_url\030\001 \001(\t\022\r\n\005value\030\002" +
-      " \001(\014Bo\n\023com.google.protobufB\010AnyProtoP\001Z" +
-      "%github.com/golang/protobuf/ptypes/any\242\002" +
-      "\003GPB\252\002\036Google.Protobuf.WellKnownTypesb\006p" +
-      "roto3"
-    };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
-          public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors(
-              org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) {
-            descriptor = root;
-            return null;
-          }
-        };
-    org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] {
-        }, assigner);
-    internal_static_google_protobuf_Any_descriptor =
-      getDescriptor().getMessageTypes().get(0);
-    internal_static_google_protobuf_Any_fieldAccessorTable = new
-      org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
-        internal_static_google_protobuf_Any_descriptor,
-        new java.lang.String[] { "TypeUrl", "Value", });
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}