You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2016/09/30 00:30:46 UTC

[17/29] hbase git commit: HBASE-16567 Upgrade to protobuf-3.1.x Regenerate all protos in this module with protoc3. Redo ByteStringer to use new pb3.1.0 unsafebytesutil instead of HBaseZeroCopyByteString

http://git-wip-us.apache.org/repos/asf/hbase/blob/b4a729ed/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
----------------------------------------------------------------------
diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
index 74d3f86..5b456c8 100644
--- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
+++ b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/protobuf/generated/FSProtos.java
@@ -6,12 +6,18 @@ package org.apache.hadoop.hbase.shaded.protobuf.generated;
 public final class FSProtos {
   private FSProtos() {}
   public static void registerAllExtensions(
+      com.google.protobuf.ExtensionRegistryLite registry) {
+  }
+
+  public static void registerAllExtensions(
       com.google.protobuf.ExtensionRegistry registry) {
+    registerAllExtensions(
+        (com.google.protobuf.ExtensionRegistryLite) registry);
   }
-  public interface HBaseVersionFileContentOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface HBaseVersionFileContentOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:hbase.pb.HBaseVersionFileContent)
+      com.google.protobuf.MessageOrBuilder {
 
-    // required string version = 1;
     /**
      * <code>required string version = 1;</code>
      */
@@ -27,43 +33,35 @@ public final class FSProtos {
         getVersionBytes();
   }
   /**
-   * Protobuf type {@code hbase.pb.HBaseVersionFileContent}
-   *
    * <pre>
    **
    * The ${HBASE_ROOTDIR}/hbase.version file content
    * </pre>
+   *
+   * Protobuf type {@code hbase.pb.HBaseVersionFileContent}
    */
-  public static final class HBaseVersionFileContent extends
-      com.google.protobuf.GeneratedMessage
-      implements HBaseVersionFileContentOrBuilder {
+  public  static final class HBaseVersionFileContent extends
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:hbase.pb.HBaseVersionFileContent)
+      HBaseVersionFileContentOrBuilder {
     // Use HBaseVersionFileContent.newBuilder() to construct.
-    private HBaseVersionFileContent(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private HBaseVersionFileContent(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
     }
-    private HBaseVersionFileContent(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final HBaseVersionFileContent defaultInstance;
-    public static HBaseVersionFileContent getDefaultInstance() {
-      return defaultInstance;
-    }
-
-    public HBaseVersionFileContent getDefaultInstanceForType() {
-      return defaultInstance;
+    private HBaseVersionFileContent() {
+      version_ = "";
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
     public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
+    getUnknownFields() {
       return this.unknownFields;
     }
     private HBaseVersionFileContent(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
+      this();
       int mutable_bitField0_ = 0;
       com.google.protobuf.UnknownFieldSet.Builder unknownFields =
           com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -83,8 +81,9 @@ public final class FSProtos {
               break;
             }
             case 10: {
+              com.google.protobuf.ByteString bs = input.readBytes();
               bitField0_ |= 0x00000001;
-              version_ = input.readBytes();
+              version_ = bs;
               break;
             }
           }
@@ -93,7 +92,7 @@ public final class FSProtos {
         throw e.setUnfinishedMessage(this);
       } catch (java.io.IOException e) {
         throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
+            e).setUnfinishedMessage(this);
       } finally {
         this.unknownFields = unknownFields.build();
         makeExtensionsImmutable();
@@ -104,32 +103,16 @@ public final class FSProtos {
       return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<HBaseVersionFileContent> PARSER =
-        new com.google.protobuf.AbstractParser<HBaseVersionFileContent>() {
-      public HBaseVersionFileContent parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new HBaseVersionFileContent(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<HBaseVersionFileContent> getParserForType() {
-      return PARSER;
-    }
-
     private int bitField0_;
-    // required string version = 1;
     public static final int VERSION_FIELD_NUMBER = 1;
-    private java.lang.Object version_;
+    private volatile java.lang.Object version_;
     /**
      * <code>required string version = 1;</code>
      */
@@ -170,13 +153,11 @@ public final class FSProtos {
       }
     }
 
-    private void initFields() {
-      version_ = "";
-    }
     private byte memoizedIsInitialized = -1;
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       if (!hasVersion()) {
         memoizedIsInitialized = 0;
@@ -188,36 +169,27 @@ public final class FSProtos {
 
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeBytes(1, getVersionBytes());
+        com.google.protobuf.GeneratedMessageV3.writeString(output, 1, version_);
       }
-      getUnknownFields().writeTo(output);
+      unknownFields.writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(1, getVersionBytes());
+        size += com.google.protobuf.GeneratedMessageV3.computeStringSize(1, version_);
       }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      size += unknownFields.getSerializedSize();
+      memoizedSize = size;
       return size;
     }
 
     private static final long serialVersionUID = 0L;
     @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
        return true;
@@ -233,12 +205,10 @@ public final class FSProtos {
         result = result && getVersion()
             .equals(other.getVersion());
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
+      result = result && unknownFields.equals(other.unknownFields);
       return result;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
@@ -250,7 +220,7 @@ public final class FSProtos {
         hash = (37 * hash) + VERSION_FIELD_NUMBER;
         hash = (53 * hash) + getVersion().hashCode();
       }
-      hash = (29 * hash) + getUnknownFields().hashCode();
+      hash = (29 * hash) + unknownFields.hashCode();
       memoizedHashCode = hash;
       return hash;
     }
@@ -278,66 +248,78 @@ public final class FSProtos {
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
     /**
-     * Protobuf type {@code hbase.pb.HBaseVersionFileContent}
-     *
      * <pre>
      **
      * The ${HBASE_ROOTDIR}/hbase.version file content
      * </pre>
+     *
+     * Protobuf type {@code hbase.pb.HBaseVersionFileContent}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContentOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:hbase.pb.HBaseVersionFileContent)
+        org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContentOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -350,18 +332,15 @@ public final class FSProtos {
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
       private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+        if (com.google.protobuf.GeneratedMessageV3
+                .alwaysUseFieldBuilders) {
         }
       }
-      private static Builder create() {
-        return new Builder();
-      }
-
       public Builder clear() {
         super.clear();
         version_ = "";
@@ -369,10 +348,6 @@ public final class FSProtos {
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
@@ -403,6 +378,32 @@ public final class FSProtos {
         return result;
       }
 
+      public Builder clone() {
+        return (Builder) super.clone();
+      }
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          Object value) {
+        return (Builder) super.setField(field, value);
+      }
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return (Builder) super.clearField(field);
+      }
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return (Builder) super.clearOneof(oneof);
+      }
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, Object value) {
+        return (Builder) super.setRepeatedField(field, index, value);
+      }
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          Object value) {
+        return (Builder) super.addRepeatedField(field, value);
+      }
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent) {
           return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent)other);
@@ -419,13 +420,13 @@ public final class FSProtos {
           version_ = other.version_;
           onChanged();
         }
-        this.mergeUnknownFields(other.getUnknownFields());
+        this.mergeUnknownFields(other.unknownFields);
+        onChanged();
         return this;
       }
 
       public final boolean isInitialized() {
         if (!hasVersion()) {
-          
           return false;
         }
         return true;
@@ -440,7 +441,7 @@ public final class FSProtos {
           parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
           parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
           if (parsedMessage != null) {
             mergeFrom(parsedMessage);
@@ -450,7 +451,6 @@ public final class FSProtos {
       }
       private int bitField0_;
 
-      // required string version = 1;
       private java.lang.Object version_ = "";
       /**
        * <code>required string version = 1;</code>
@@ -464,9 +464,12 @@ public final class FSProtos {
       public java.lang.String getVersion() {
         java.lang.Object ref = version_;
         if (!(ref instanceof java.lang.String)) {
-          java.lang.String s = ((com.google.protobuf.ByteString) ref)
-              .toStringUtf8();
-          version_ = s;
+          com.google.protobuf.ByteString bs =
+              (com.google.protobuf.ByteString) ref;
+          java.lang.String s = bs.toStringUtf8();
+          if (bs.isValidUtf8()) {
+            version_ = s;
+          }
           return s;
         } else {
           return (java.lang.String) ref;
@@ -523,22 +526,59 @@ public final class FSProtos {
         onChanged();
         return this;
       }
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:hbase.pb.HBaseVersionFileContent)
     }
 
+    // @@protoc_insertion_point(class_scope:hbase.pb.HBaseVersionFileContent)
+    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent DEFAULT_INSTANCE;
     static {
-      defaultInstance = new HBaseVersionFileContent(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent();
+    }
+
+    public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
+
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<HBaseVersionFileContent>
+        PARSER = new com.google.protobuf.AbstractParser<HBaseVersionFileContent>() {
+      public HBaseVersionFileContent parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+          return new HBaseVersionFileContent(input, extensionRegistry);
+      }
+    };
+
+    public static com.google.protobuf.Parser<HBaseVersionFileContent> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<HBaseVersionFileContent> getParserForType() {
+      return PARSER;
+    }
+
+    public org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.HBaseVersionFileContent getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
     }
 
-    // @@protoc_insertion_point(class_scope:hbase.pb.HBaseVersionFileContent)
   }
 
-  public interface ReferenceOrBuilder
-      extends com.google.protobuf.MessageOrBuilder {
+  public interface ReferenceOrBuilder extends
+      // @@protoc_insertion_point(interface_extends:hbase.pb.Reference)
+      com.google.protobuf.MessageOrBuilder {
 
-    // required bytes splitkey = 1;
     /**
      * <code>required bytes splitkey = 1;</code>
      */
@@ -548,7 +588,6 @@ public final class FSProtos {
      */
     com.google.protobuf.ByteString getSplitkey();
 
-    // required .hbase.pb.Reference.Range range = 2;
     /**
      * <code>required .hbase.pb.Reference.Range range = 2;</code>
      */
@@ -559,43 +598,36 @@ public final class FSProtos {
     org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range getRange();
   }
   /**
-   * Protobuf type {@code hbase.pb.Reference}
-   *
    * <pre>
    **
    * Reference file content used when we split an hfile under a region.
    * </pre>
+   *
+   * Protobuf type {@code hbase.pb.Reference}
    */
-  public static final class Reference extends
-      com.google.protobuf.GeneratedMessage
-      implements ReferenceOrBuilder {
+  public  static final class Reference extends
+      com.google.protobuf.GeneratedMessageV3 implements
+      // @@protoc_insertion_point(message_implements:hbase.pb.Reference)
+      ReferenceOrBuilder {
     // Use Reference.newBuilder() to construct.
-    private Reference(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+    private Reference(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
-      this.unknownFields = builder.getUnknownFields();
-    }
-    private Reference(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
-
-    private static final Reference defaultInstance;
-    public static Reference getDefaultInstance() {
-      return defaultInstance;
     }
-
-    public Reference getDefaultInstanceForType() {
-      return defaultInstance;
+    private Reference() {
+      splitkey_ = com.google.protobuf.ByteString.EMPTY;
+      range_ = 0;
     }
 
-    private final com.google.protobuf.UnknownFieldSet unknownFields;
     @java.lang.Override
     public final com.google.protobuf.UnknownFieldSet
-        getUnknownFields() {
+    getUnknownFields() {
       return this.unknownFields;
     }
     private Reference(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      initFields();
+      this();
       int mutable_bitField0_ = 0;
       com.google.protobuf.UnknownFieldSet.Builder unknownFields =
           com.google.protobuf.UnknownFieldSet.newBuilder();
@@ -626,7 +658,7 @@ public final class FSProtos {
                 unknownFields.mergeVarintField(2, rawValue);
               } else {
                 bitField0_ |= 0x00000002;
-                range_ = value;
+                range_ = rawValue;
               }
               break;
             }
@@ -636,7 +668,7 @@ public final class FSProtos {
         throw e.setUnfinishedMessage(this);
       } catch (java.io.IOException e) {
         throw new com.google.protobuf.InvalidProtocolBufferException(
-            e.getMessage()).setUnfinishedMessage(this);
+            e).setUnfinishedMessage(this);
       } finally {
         this.unknownFields = unknownFields.build();
         makeExtensionsImmutable();
@@ -647,28 +679,13 @@ public final class FSProtos {
       return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
     }
 
-    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+    protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
         internalGetFieldAccessorTable() {
       return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
               org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.class, org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<Reference> PARSER =
-        new com.google.protobuf.AbstractParser<Reference>() {
-      public Reference parsePartialFrom(
-          com.google.protobuf.CodedInputStream input,
-          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        return new Reference(input, extensionRegistry);
-      }
-    };
-
-    @java.lang.Override
-    public com.google.protobuf.Parser<Reference> getParserForType() {
-      return PARSER;
-    }
-
     /**
      * Protobuf enum {@code hbase.pb.Reference.Range}
      */
@@ -677,11 +694,11 @@ public final class FSProtos {
       /**
        * <code>TOP = 0;</code>
        */
-      TOP(0, 0),
+      TOP(0),
       /**
        * <code>BOTTOM = 1;</code>
        */
-      BOTTOM(1, 1),
+      BOTTOM(1),
       ;
 
       /**
@@ -694,9 +711,19 @@ public final class FSProtos {
       public static final int BOTTOM_VALUE = 1;
 
 
-      public final int getNumber() { return value; }
+      public final int getNumber() {
+        return value;
+      }
 
+      /**
+       * @deprecated Use {@link #forNumber(int)} instead.
+       */
+      @java.lang.Deprecated
       public static Range valueOf(int value) {
+        return forNumber(value);
+      }
+
+      public static Range forNumber(int value) {
         switch (value) {
           case 0: return TOP;
           case 1: return BOTTOM;
@@ -708,17 +735,17 @@ public final class FSProtos {
           internalGetValueMap() {
         return internalValueMap;
       }
-      private static com.google.protobuf.Internal.EnumLiteMap<Range>
-          internalValueMap =
+      private static final com.google.protobuf.Internal.EnumLiteMap<
+          Range> internalValueMap =
             new com.google.protobuf.Internal.EnumLiteMap<Range>() {
               public Range findValueByNumber(int number) {
-                return Range.valueOf(number);
+                return Range.forNumber(number);
               }
             };
 
       public final com.google.protobuf.Descriptors.EnumValueDescriptor
           getValueDescriptor() {
-        return getDescriptor().getValues().get(index);
+        return getDescriptor().getValues().get(ordinal());
       }
       public final com.google.protobuf.Descriptors.EnumDescriptor
           getDescriptorForType() {
@@ -740,11 +767,9 @@ public final class FSProtos {
         return VALUES[desc.getIndex()];
       }
 
-      private final int index;
       private final int value;
 
-      private Range(int index, int value) {
-        this.index = index;
+      private Range(int value) {
         this.value = value;
       }
 
@@ -752,7 +777,6 @@ public final class FSProtos {
     }
 
     private int bitField0_;
-    // required bytes splitkey = 1;
     public static final int SPLITKEY_FIELD_NUMBER = 1;
     private com.google.protobuf.ByteString splitkey_;
     /**
@@ -768,9 +792,8 @@ public final class FSProtos {
       return splitkey_;
     }
 
-    // required .hbase.pb.Reference.Range range = 2;
     public static final int RANGE_FIELD_NUMBER = 2;
-    private org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range range_;
+    private int range_;
     /**
      * <code>required .hbase.pb.Reference.Range range = 2;</code>
      */
@@ -781,17 +804,15 @@ public final class FSProtos {
      * <code>required .hbase.pb.Reference.Range range = 2;</code>
      */
     public org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range getRange() {
-      return range_;
+      org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range result = org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.valueOf(range_);
+      return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP : result;
     }
 
-    private void initFields() {
-      splitkey_ = com.google.protobuf.ByteString.EMPTY;
-      range_ = org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP;
-    }
     private byte memoizedIsInitialized = -1;
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
-      if (isInitialized != -1) return isInitialized == 1;
+      if (isInitialized == 1) return true;
+      if (isInitialized == 0) return false;
 
       if (!hasSplitkey()) {
         memoizedIsInitialized = 0;
@@ -807,19 +828,17 @@ public final class FSProtos {
 
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
-      getSerializedSize();
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
         output.writeBytes(1, splitkey_);
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeEnum(2, range_.getNumber());
+        output.writeEnum(2, range_);
       }
-      getUnknownFields().writeTo(output);
+      unknownFields.writeTo(output);
     }
 
-    private int memoizedSerializedSize = -1;
     public int getSerializedSize() {
-      int size = memoizedSerializedSize;
+      int size = memoizedSize;
       if (size != -1) return size;
 
       size = 0;
@@ -829,21 +848,15 @@ public final class FSProtos {
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeEnumSize(2, range_.getNumber());
+          .computeEnumSize(2, range_);
       }
-      size += getUnknownFields().getSerializedSize();
-      memoizedSerializedSize = size;
+      size += unknownFields.getSerializedSize();
+      memoizedSize = size;
       return size;
     }
 
     private static final long serialVersionUID = 0L;
     @java.lang.Override
-    protected java.lang.Object writeReplace()
-        throws java.io.ObjectStreamException {
-      return super.writeReplace();
-    }
-
-    @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
        return true;
@@ -861,15 +874,12 @@ public final class FSProtos {
       }
       result = result && (hasRange() == other.hasRange());
       if (hasRange()) {
-        result = result &&
-            (getRange() == other.getRange());
+        result = result && range_ == other.range_;
       }
-      result = result &&
-          getUnknownFields().equals(other.getUnknownFields());
+      result = result && unknownFields.equals(other.unknownFields);
       return result;
     }
 
-    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
       if (memoizedHashCode != 0) {
@@ -883,9 +893,9 @@ public final class FSProtos {
       }
       if (hasRange()) {
         hash = (37 * hash) + RANGE_FIELD_NUMBER;
-        hash = (53 * hash) + hashEnum(getRange());
+        hash = (53 * hash) + range_;
       }
-      hash = (29 * hash) + getUnknownFields().hashCode();
+      hash = (29 * hash) + unknownFields.hashCode();
       memoizedHashCode = hash;
       return hash;
     }
@@ -913,66 +923,78 @@ public final class FSProtos {
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return PARSER.parseFrom(input);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input);
     }
     public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return PARSER.parseFrom(input, extensionRegistry);
+      return com.google.protobuf.GeneratedMessageV3
+          .parseWithIOException(PARSER, input, extensionRegistry);
     }
 
-    public static Builder newBuilder() { return Builder.create(); }
     public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder() {
+      return DEFAULT_INSTANCE.toBuilder();
+    }
     public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference prototype) {
-      return newBuilder().mergeFrom(prototype);
+      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
+    }
+    public Builder toBuilder() {
+      return this == DEFAULT_INSTANCE
+          ? new Builder() : new Builder().mergeFrom(this);
     }
-    public Builder toBuilder() { return newBuilder(this); }
 
     @java.lang.Override
     protected Builder newBuilderForType(
-        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
     /**
-     * Protobuf type {@code hbase.pb.Reference}
-     *
      * <pre>
      **
      * Reference file content used when we split an hfile under a region.
      * </pre>
+     *
+     * Protobuf type {@code hbase.pb.Reference}
      */
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder {
+        com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements
+        // @@protoc_insertion_point(builder_implements:hbase.pb.Reference)
+        org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.ReferenceOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
         return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
       }
 
-      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      protected com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
           internalGetFieldAccessorTable() {
         return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
@@ -985,31 +1007,24 @@ public final class FSProtos {
       }
 
       private Builder(
-          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+          com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
       private void maybeForceBuilderInitialization() {
-        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+        if (com.google.protobuf.GeneratedMessageV3
+                .alwaysUseFieldBuilders) {
         }
       }
-      private static Builder create() {
-        return new Builder();
-      }
-
       public Builder clear() {
         super.clear();
         splitkey_ = com.google.protobuf.ByteString.EMPTY;
         bitField0_ = (bitField0_ & ~0x00000001);
-        range_ = org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP;
+        range_ = 0;
         bitField0_ = (bitField0_ & ~0x00000002);
         return this;
       }
 
-      public Builder clone() {
-        return create().mergeFrom(buildPartial());
-      }
-
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
         return org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.internal_static_hbase_pb_Reference_descriptor;
@@ -1044,6 +1059,32 @@ public final class FSProtos {
         return result;
       }
 
+      public Builder clone() {
+        return (Builder) super.clone();
+      }
+      public Builder setField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          Object value) {
+        return (Builder) super.setField(field, value);
+      }
+      public Builder clearField(
+          com.google.protobuf.Descriptors.FieldDescriptor field) {
+        return (Builder) super.clearField(field);
+      }
+      public Builder clearOneof(
+          com.google.protobuf.Descriptors.OneofDescriptor oneof) {
+        return (Builder) super.clearOneof(oneof);
+      }
+      public Builder setRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          int index, Object value) {
+        return (Builder) super.setRepeatedField(field, index, value);
+      }
+      public Builder addRepeatedField(
+          com.google.protobuf.Descriptors.FieldDescriptor field,
+          Object value) {
+        return (Builder) super.addRepeatedField(field, value);
+      }
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference) {
           return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference)other);
@@ -1061,17 +1102,16 @@ public final class FSProtos {
         if (other.hasRange()) {
           setRange(other.getRange());
         }
-        this.mergeUnknownFields(other.getUnknownFields());
+        this.mergeUnknownFields(other.unknownFields);
+        onChanged();
         return this;
       }
 
       public final boolean isInitialized() {
         if (!hasSplitkey()) {
-          
           return false;
         }
         if (!hasRange()) {
-          
           return false;
         }
         return true;
@@ -1086,7 +1126,7 @@ public final class FSProtos {
           parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
           parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference) e.getUnfinishedMessage();
-          throw e;
+          throw e.unwrapIOException();
         } finally {
           if (parsedMessage != null) {
             mergeFrom(parsedMessage);
@@ -1096,7 +1136,6 @@ public final class FSProtos {
       }
       private int bitField0_;
 
-      // required bytes splitkey = 1;
       private com.google.protobuf.ByteString splitkey_ = com.google.protobuf.ByteString.EMPTY;
       /**
        * <code>required bytes splitkey = 1;</code>
@@ -1132,8 +1171,7 @@ public final class FSProtos {
         return this;
       }
 
-      // required .hbase.pb.Reference.Range range = 2;
-      private org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range range_ = org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP;
+      private int range_ = 0;
       /**
        * <code>required .hbase.pb.Reference.Range range = 2;</code>
        */
@@ -1144,7 +1182,8 @@ public final class FSProtos {
        * <code>required .hbase.pb.Reference.Range range = 2;</code>
        */
       public org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range getRange() {
-        return range_;
+        org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range result = org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.valueOf(range_);
+        return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP : result;
       }
       /**
        * <code>required .hbase.pb.Reference.Range range = 2;</code>
@@ -1154,7 +1193,7 @@ public final class FSProtos {
           throw new NullPointerException();
         }
         bitField0_ |= 0x00000002;
-        range_ = value;
+        range_ = value.getNumber();
         onChanged();
         return this;
       }
@@ -1163,38 +1202,75 @@ public final class FSProtos {
        */
       public Builder clearRange() {
         bitField0_ = (bitField0_ & ~0x00000002);
-        range_ = org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference.Range.TOP;
+        range_ = 0;
         onChanged();
         return this;
       }
+      public final Builder setUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.setUnknownFields(unknownFields);
+      }
+
+      public final Builder mergeUnknownFields(
+          final com.google.protobuf.UnknownFieldSet unknownFields) {
+        return super.mergeUnknownFields(unknownFields);
+      }
+
 
       // @@protoc_insertion_point(builder_scope:hbase.pb.Reference)
     }
 
+    // @@protoc_insertion_point(class_scope:hbase.pb.Reference)
+    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference DEFAULT_INSTANCE;
     static {
-      defaultInstance = new Reference(true);
-      defaultInstance.initFields();
+      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference();
+    }
+
+    public static org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference getDefaultInstance() {
+      return DEFAULT_INSTANCE;
+    }
+
+    @java.lang.Deprecated public static final com.google.protobuf.Parser<Reference>
+        PARSER = new com.google.protobuf.AbstractParser<Reference>() {
+      public Reference parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+          return new Reference(input, extensionRegistry);
+      }
+    };
+
+    public static com.google.protobuf.Parser<Reference> parser() {
+      return PARSER;
+    }
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<Reference> getParserForType() {
+      return PARSER;
+    }
+
+    public org.apache.hadoop.hbase.shaded.protobuf.generated.FSProtos.Reference getDefaultInstanceForType() {
+      return DEFAULT_INSTANCE;
     }
 
-    // @@protoc_insertion_point(class_scope:hbase.pb.Reference)
   }
 
-  private static com.google.protobuf.Descriptors.Descriptor
+  private static final com.google.protobuf.Descriptors.Descriptor
     internal_static_hbase_pb_HBaseVersionFileContent_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
+  private static final 
+    com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
       internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable;
-  private static com.google.protobuf.Descriptors.Descriptor
+  private static final com.google.protobuf.Descriptors.Descriptor
     internal_static_hbase_pb_Reference_descriptor;
-  private static
-    com.google.protobuf.GeneratedMessage.FieldAccessorTable
+  private static final 
+    com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
       internal_static_hbase_pb_Reference_fieldAccessorTable;
 
   public static com.google.protobuf.Descriptors.FileDescriptor
       getDescriptor() {
     return descriptor;
   }
-  private static com.google.protobuf.Descriptors.FileDescriptor
+  private static  com.google.protobuf.Descriptors.FileDescriptor
       descriptor;
   static {
     java.lang.String[] descriptorData = {
@@ -1206,29 +1282,29 @@ public final class FSProtos {
       "d.protobuf.generatedB\010FSProtosH\001\240\001\001"
     };
     com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          internal_static_hbase_pb_HBaseVersionFileContent_descriptor =
-            getDescriptor().getMessageTypes().get(0);
-          internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_pb_HBaseVersionFileContent_descriptor,
-              new java.lang.String[] { "Version", });
-          internal_static_hbase_pb_Reference_descriptor =
-            getDescriptor().getMessageTypes().get(1);
-          internal_static_hbase_pb_Reference_fieldAccessorTable = new
-            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
-              internal_static_hbase_pb_Reference_descriptor,
-              new java.lang.String[] { "Splitkey", "Range", });
-          return null;
-        }
-      };
+        new com.google.protobuf.Descriptors.FileDescriptor.    InternalDescriptorAssigner() {
+          public com.google.protobuf.ExtensionRegistry assignDescriptors(
+              com.google.protobuf.Descriptors.FileDescriptor root) {
+            descriptor = root;
+            return null;
+          }
+        };
     com.google.protobuf.Descriptors.FileDescriptor
       .internalBuildGeneratedFileFrom(descriptorData,
         new com.google.protobuf.Descriptors.FileDescriptor[] {
         }, assigner);
+    internal_static_hbase_pb_HBaseVersionFileContent_descriptor =
+      getDescriptor().getMessageTypes().get(0);
+    internal_static_hbase_pb_HBaseVersionFileContent_fieldAccessorTable = new
+      com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+        internal_static_hbase_pb_HBaseVersionFileContent_descriptor,
+        new java.lang.String[] { "Version", });
+    internal_static_hbase_pb_Reference_descriptor =
+      getDescriptor().getMessageTypes().get(1);
+    internal_static_hbase_pb_Reference_fieldAccessorTable = new
+      com.google.protobuf.GeneratedMessageV3.FieldAccessorTable(
+        internal_static_hbase_pb_Reference_descriptor,
+        new java.lang.String[] { "Splitkey", "Range", });
   }
 
   // @@protoc_insertion_point(outer_class_scope)