You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/08/21 07:10:45 UTC

svn commit: r1516086 [27/43] - in /hbase/branches/0.95: ./ hbase-client/src/main/java/org/apache/hadoop/hbase/ hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ hbase-common/src/test/java/org/apache/hadoop/hbase/ hbase-protocol/src/main/java...

Modified: hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java?rev=1516086&r1=1516085&r2=1516086&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java (original)
+++ hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java Wed Aug 21 05:10:43 2013
@@ -8,18 +8,33 @@ public final class WALProtos {
   public static void registerAllExtensions(
       com.google.protobuf.ExtensionRegistry registry) {
   }
+  /**
+   * Protobuf enum {@code ScopeType}
+   */
   public enum ScopeType
       implements com.google.protobuf.ProtocolMessageEnum {
+    /**
+     * <code>REPLICATION_SCOPE_LOCAL = 0;</code>
+     */
     REPLICATION_SCOPE_LOCAL(0, 0),
+    /**
+     * <code>REPLICATION_SCOPE_GLOBAL = 1;</code>
+     */
     REPLICATION_SCOPE_GLOBAL(1, 1),
     ;
-    
+
+    /**
+     * <code>REPLICATION_SCOPE_LOCAL = 0;</code>
+     */
     public static final int REPLICATION_SCOPE_LOCAL_VALUE = 0;
+    /**
+     * <code>REPLICATION_SCOPE_GLOBAL = 1;</code>
+     */
     public static final int REPLICATION_SCOPE_GLOBAL_VALUE = 1;
-    
-    
+
+
     public final int getNumber() { return value; }
-    
+
     public static ScopeType valueOf(int value) {
       switch (value) {
         case 0: return REPLICATION_SCOPE_LOCAL;
@@ -27,7 +42,7 @@ public final class WALProtos {
         default: return null;
       }
     }
-    
+
     public static com.google.protobuf.Internal.EnumLiteMap<ScopeType>
         internalGetValueMap() {
       return internalValueMap;
@@ -39,7 +54,7 @@ public final class WALProtos {
               return ScopeType.valueOf(number);
             }
           };
-    
+
     public final com.google.protobuf.Descriptors.EnumValueDescriptor
         getValueDescriptor() {
       return getDescriptor().getValues().get(index);
@@ -52,11 +67,9 @@ public final class WALProtos {
         getDescriptor() {
       return org.apache.hadoop.hbase.protobuf.generated.WALProtos.getDescriptor().getEnumTypes().get(0);
     }
-    
-    private static final ScopeType[] VALUES = {
-      REPLICATION_SCOPE_LOCAL, REPLICATION_SCOPE_GLOBAL, 
-    };
-    
+
+    private static final ScopeType[] VALUES = values();
+
     public static ScopeType valueOf(
         com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
       if (desc.getType() != getDescriptor()) {
@@ -65,64 +78,143 @@ public final class WALProtos {
       }
       return VALUES[desc.getIndex()];
     }
-    
+
     private final int index;
     private final int value;
-    
+
     private ScopeType(int index, int value) {
       this.index = index;
       this.value = value;
     }
-    
+
     // @@protoc_insertion_point(enum_scope:ScopeType)
   }
-  
+
   public interface WALHeaderOrBuilder
       extends com.google.protobuf.MessageOrBuilder {
-    
+
     // optional bool has_compression = 1;
+    /**
+     * <code>optional bool has_compression = 1;</code>
+     */
     boolean hasHasCompression();
+    /**
+     * <code>optional bool has_compression = 1;</code>
+     */
     boolean getHasCompression();
   }
+  /**
+   * Protobuf type {@code WALHeader}
+   */
   public static final class WALHeader extends
       com.google.protobuf.GeneratedMessage
       implements WALHeaderOrBuilder {
     // Use WALHeader.newBuilder() to construct.
-    private WALHeader(Builder builder) {
+    private WALHeader(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
       super(builder);
+      this.unknownFields = builder.getUnknownFields();
     }
-    private WALHeader(boolean noInit) {}
-    
+    private WALHeader(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
     private static final WALHeader defaultInstance;
     public static WALHeader getDefaultInstance() {
       return defaultInstance;
     }
-    
+
     public WALHeader getDefaultInstanceForType() {
       return defaultInstance;
     }
-    
+
+    private final com.google.protobuf.UnknownFieldSet unknownFields;
+    @java.lang.Override
+    public final com.google.protobuf.UnknownFieldSet
+        getUnknownFields() {
+      return this.unknownFields;
+    }
+    private WALHeader(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      initFields();
+      int mutable_bitField0_ = 0;
+      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+          com.google.protobuf.UnknownFieldSet.newBuilder();
+      try {
+        boolean done = false;
+        while (!done) {
+          int tag = input.readTag();
+          switch (tag) {
+            case 0:
+              done = true;
+              break;
+            default: {
+              if (!parseUnknownField(input, unknownFields,
+                                     extensionRegistry, tag)) {
+                done = true;
+              }
+              break;
+            }
+            case 8: {
+              bitField0_ |= 0x00000001;
+              hasCompression_ = input.readBool();
+              break;
+            }
+          }
+        }
+      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+        throw e.setUnfinishedMessage(this);
+      } catch (java.io.IOException e) {
+        throw new com.google.protobuf.InvalidProtocolBufferException(
+            e.getMessage()).setUnfinishedMessage(this);
+      } finally {
+        this.unknownFields = unknownFields.build();
+        makeExtensionsImmutable();
+      }
+    }
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALHeader_descriptor;
     }
-    
+
     protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
         internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALHeader_fieldAccessorTable;
+      return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALHeader_fieldAccessorTable
+          .ensureFieldAccessorsInitialized(
+              org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader.class, org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader.Builder.class);
+    }
+
+    public static com.google.protobuf.Parser<WALHeader> PARSER =
+        new com.google.protobuf.AbstractParser<WALHeader>() {
+      public WALHeader parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        return new WALHeader(input, extensionRegistry);
+      }
+    };
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<WALHeader> getParserForType() {
+      return PARSER;
     }
-    
+
     private int bitField0_;
     // optional bool has_compression = 1;
     public static final int HAS_COMPRESSION_FIELD_NUMBER = 1;
     private boolean hasCompression_;
+    /**
+     * <code>optional bool has_compression = 1;</code>
+     */
     public boolean hasHasCompression() {
       return ((bitField0_ & 0x00000001) == 0x00000001);
     }
+    /**
+     * <code>optional bool has_compression = 1;</code>
+     */
     public boolean getHasCompression() {
       return hasCompression_;
     }
-    
+
     private void initFields() {
       hasCompression_ = false;
     }
@@ -130,11 +222,11 @@ public final class WALProtos {
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
       if (isInitialized != -1) return isInitialized == 1;
-      
+
       memoizedIsInitialized = 1;
       return true;
     }
-    
+
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
       getSerializedSize();
@@ -143,12 +235,12 @@ public final class WALProtos {
       }
       getUnknownFields().writeTo(output);
     }
-    
+
     private int memoizedSerializedSize = -1;
     public int getSerializedSize() {
       int size = memoizedSerializedSize;
       if (size != -1) return size;
-    
+
       size = 0;
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
         size += com.google.protobuf.CodedOutputStream
@@ -158,14 +250,14 @@ public final class WALProtos {
       memoizedSerializedSize = size;
       return size;
     }
-    
+
     private static final long serialVersionUID = 0L;
     @java.lang.Override
     protected java.lang.Object writeReplace()
         throws java.io.ObjectStreamException {
       return super.writeReplace();
     }
-    
+
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -175,7 +267,7 @@ public final class WALProtos {
         return super.equals(obj);
       }
       org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader other = (org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader) obj;
-      
+
       boolean result = true;
       result = result && (hasHasCompression() == other.hasHasCompression());
       if (hasHasCompression()) {
@@ -186,9 +278,13 @@ public final class WALProtos {
           getUnknownFields().equals(other.getUnknownFields());
       return result;
     }
-    
+
+    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
+      if (memoizedHashCode != 0) {
+        return memoizedHashCode;
+      }
       int hash = 41;
       hash = (19 * hash) + getDescriptorForType().hashCode();
       if (hasHasCompression()) {
@@ -196,89 +292,79 @@ public final class WALProtos {
         hash = (53 * hash) + hashBoolean(getHasCompression());
       }
       hash = (29 * hash) + getUnknownFields().hashCode();
+      memoizedHashCode = hash;
       return hash;
     }
-    
+
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+      return PARSER.parseFrom(data);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseFrom(
         com.google.protobuf.ByteString data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(data, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseFrom(byte[] data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+      return PARSER.parseFrom(data);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseFrom(
         byte[] data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(data, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+      return PARSER.parseFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseDelimitedFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseDelimitedFrom(input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+      return PARSER.parseFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(input, extensionRegistry);
     }
-    
+
     public static Builder newBuilder() { return Builder.create(); }
     public Builder newBuilderForType() { return newBuilder(); }
     public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader prototype) {
       return newBuilder().mergeFrom(prototype);
     }
     public Builder toBuilder() { return newBuilder(this); }
-    
+
     @java.lang.Override
     protected Builder newBuilderForType(
         com.google.protobuf.GeneratedMessage.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
+    /**
+     * Protobuf type {@code WALHeader}
+     */
     public static final class Builder extends
         com.google.protobuf.GeneratedMessage.Builder<Builder>
        implements org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeaderOrBuilder {
@@ -286,18 +372,21 @@ public final class WALProtos {
           getDescriptor() {
         return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALHeader_descriptor;
       }
-      
+
       protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
           internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALHeader_fieldAccessorTable;
+        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALHeader_fieldAccessorTable
+            .ensureFieldAccessorsInitialized(
+                org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader.class, org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader.Builder.class);
       }
-      
+
       // Construct using org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader.newBuilder()
       private Builder() {
         maybeForceBuilderInitialization();
       }
-      
-      private Builder(BuilderParent parent) {
+
+      private Builder(
+          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
@@ -308,27 +397,27 @@ public final class WALProtos {
       private static Builder create() {
         return new Builder();
       }
-      
+
       public Builder clear() {
         super.clear();
         hasCompression_ = false;
         bitField0_ = (bitField0_ & ~0x00000001);
         return this;
       }
-      
+
       public Builder clone() {
         return create().mergeFrom(buildPartial());
       }
-      
+
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader.getDescriptor();
+        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALHeader_descriptor;
       }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader getDefaultInstanceForType() {
         return org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader.getDefaultInstance();
       }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader build() {
         org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader result = buildPartial();
         if (!result.isInitialized()) {
@@ -336,17 +425,7 @@ public final class WALProtos {
         }
         return result;
       }
-      
-      private org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader buildParsed()
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(
-            result).asInvalidProtocolBufferException();
-        }
-        return result;
-      }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader buildPartial() {
         org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader result = new org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader(this);
         int from_bitField0_ = bitField0_;
@@ -359,7 +438,7 @@ public final class WALProtos {
         onBuilt();
         return result;
       }
-      
+
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader) {
           return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader)other);
@@ -368,7 +447,7 @@ public final class WALProtos {
           return this;
         }
       }
-      
+
       public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader other) {
         if (other == org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader.getDefaultInstance()) return this;
         if (other.hasHasCompression()) {
@@ -377,228 +456,484 @@ public final class WALProtos {
         this.mergeUnknownFields(other.getUnknownFields());
         return this;
       }
-      
+
       public final boolean isInitialized() {
         return true;
       }
-      
+
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder(
-            this.getUnknownFields());
-        while (true) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              this.setUnknownFields(unknownFields.build());
-              onChanged();
-              return this;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                this.setUnknownFields(unknownFields.build());
-                onChanged();
-                return this;
-              }
-              break;
-            }
-            case 8: {
-              bitField0_ |= 0x00000001;
-              hasCompression_ = input.readBool();
-              break;
-            }
+        org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader parsedMessage = null;
+        try {
+          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALHeader) e.getUnfinishedMessage();
+          throw e;
+        } finally {
+          if (parsedMessage != null) {
+            mergeFrom(parsedMessage);
           }
         }
+        return this;
       }
-      
       private int bitField0_;
-      
+
       // optional bool has_compression = 1;
       private boolean hasCompression_ ;
+      /**
+       * <code>optional bool has_compression = 1;</code>
+       */
       public boolean hasHasCompression() {
         return ((bitField0_ & 0x00000001) == 0x00000001);
       }
+      /**
+       * <code>optional bool has_compression = 1;</code>
+       */
       public boolean getHasCompression() {
         return hasCompression_;
       }
+      /**
+       * <code>optional bool has_compression = 1;</code>
+       */
       public Builder setHasCompression(boolean value) {
         bitField0_ |= 0x00000001;
         hasCompression_ = value;
         onChanged();
         return this;
       }
+      /**
+       * <code>optional bool has_compression = 1;</code>
+       */
       public Builder clearHasCompression() {
         bitField0_ = (bitField0_ & ~0x00000001);
         hasCompression_ = false;
         onChanged();
         return this;
       }
-      
+
       // @@protoc_insertion_point(builder_scope:WALHeader)
     }
-    
+
     static {
       defaultInstance = new WALHeader(true);
       defaultInstance.initFields();
     }
-    
+
     // @@protoc_insertion_point(class_scope:WALHeader)
   }
-  
+
   public interface WALKeyOrBuilder
       extends com.google.protobuf.MessageOrBuilder {
-    
+
     // required bytes encoded_region_name = 1;
+    /**
+     * <code>required bytes encoded_region_name = 1;</code>
+     */
     boolean hasEncodedRegionName();
+    /**
+     * <code>required bytes encoded_region_name = 1;</code>
+     */
     com.google.protobuf.ByteString getEncodedRegionName();
-    
+
     // required bytes table_name = 2;
+    /**
+     * <code>required bytes table_name = 2;</code>
+     */
     boolean hasTableName();
+    /**
+     * <code>required bytes table_name = 2;</code>
+     */
     com.google.protobuf.ByteString getTableName();
-    
+
     // required uint64 log_sequence_number = 3;
+    /**
+     * <code>required uint64 log_sequence_number = 3;</code>
+     */
     boolean hasLogSequenceNumber();
+    /**
+     * <code>required uint64 log_sequence_number = 3;</code>
+     */
     long getLogSequenceNumber();
-    
+
     // required uint64 write_time = 4;
+    /**
+     * <code>required uint64 write_time = 4;</code>
+     */
     boolean hasWriteTime();
+    /**
+     * <code>required uint64 write_time = 4;</code>
+     */
     long getWriteTime();
-    
+
     // optional .UUID cluster_id = 5;
+    /**
+     * <code>optional .UUID cluster_id = 5;</code>
+     */
     boolean hasClusterId();
+    /**
+     * <code>optional .UUID cluster_id = 5;</code>
+     */
     org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId();
+    /**
+     * <code>optional .UUID cluster_id = 5;</code>
+     */
     org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder();
-    
+
     // repeated .FamilyScope scopes = 6;
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     java.util.List<org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope> 
         getScopesList();
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope getScopes(int index);
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     int getScopesCount();
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder> 
         getScopesOrBuilderList();
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder(
         int index);
-    
+
     // optional uint32 following_kv_count = 7;
+    /**
+     * <code>optional uint32 following_kv_count = 7;</code>
+     *
+     * <pre>
+     *
+     *optional CustomEntryType custom_entry_type = 8;
+     *
+     *enum CustomEntryType {
+     *COMPACTION = 0;
+     *}
+     * </pre>
+     */
     boolean hasFollowingKvCount();
+    /**
+     * <code>optional uint32 following_kv_count = 7;</code>
+     *
+     * <pre>
+     *
+     *optional CustomEntryType custom_entry_type = 8;
+     *
+     *enum CustomEntryType {
+     *COMPACTION = 0;
+     *}
+     * </pre>
+     */
     int getFollowingKvCount();
   }
+  /**
+   * Protobuf type {@code WALKey}
+   *
+   * <pre>
+   * Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs
+   * </pre>
+   */
   public static final class WALKey extends
       com.google.protobuf.GeneratedMessage
       implements WALKeyOrBuilder {
     // Use WALKey.newBuilder() to construct.
-    private WALKey(Builder builder) {
+    private WALKey(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
       super(builder);
+      this.unknownFields = builder.getUnknownFields();
     }
-    private WALKey(boolean noInit) {}
-    
+    private WALKey(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
     private static final WALKey defaultInstance;
     public static WALKey getDefaultInstance() {
       return defaultInstance;
     }
-    
+
     public WALKey getDefaultInstanceForType() {
       return defaultInstance;
     }
-    
+
+    private final com.google.protobuf.UnknownFieldSet unknownFields;
+    @java.lang.Override
+    public final com.google.protobuf.UnknownFieldSet
+        getUnknownFields() {
+      return this.unknownFields;
+    }
+    private WALKey(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      initFields();
+      int mutable_bitField0_ = 0;
+      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+          com.google.protobuf.UnknownFieldSet.newBuilder();
+      try {
+        boolean done = false;
+        while (!done) {
+          int tag = input.readTag();
+          switch (tag) {
+            case 0:
+              done = true;
+              break;
+            default: {
+              if (!parseUnknownField(input, unknownFields,
+                                     extensionRegistry, tag)) {
+                done = true;
+              }
+              break;
+            }
+            case 10: {
+              bitField0_ |= 0x00000001;
+              encodedRegionName_ = input.readBytes();
+              break;
+            }
+            case 18: {
+              bitField0_ |= 0x00000002;
+              tableName_ = input.readBytes();
+              break;
+            }
+            case 24: {
+              bitField0_ |= 0x00000004;
+              logSequenceNumber_ = input.readUInt64();
+              break;
+            }
+            case 32: {
+              bitField0_ |= 0x00000008;
+              writeTime_ = input.readUInt64();
+              break;
+            }
+            case 42: {
+              org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder subBuilder = null;
+              if (((bitField0_ & 0x00000010) == 0x00000010)) {
+                subBuilder = clusterId_.toBuilder();
+              }
+              clusterId_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.PARSER, extensionRegistry);
+              if (subBuilder != null) {
+                subBuilder.mergeFrom(clusterId_);
+                clusterId_ = subBuilder.buildPartial();
+              }
+              bitField0_ |= 0x00000010;
+              break;
+            }
+            case 50: {
+              if (!((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
+                scopes_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope>();
+                mutable_bitField0_ |= 0x00000020;
+              }
+              scopes_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.PARSER, extensionRegistry));
+              break;
+            }
+            case 56: {
+              bitField0_ |= 0x00000020;
+              followingKvCount_ = input.readUInt32();
+              break;
+            }
+          }
+        }
+      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+        throw e.setUnfinishedMessage(this);
+      } catch (java.io.IOException e) {
+        throw new com.google.protobuf.InvalidProtocolBufferException(
+            e.getMessage()).setUnfinishedMessage(this);
+      } finally {
+        if (((mutable_bitField0_ & 0x00000020) == 0x00000020)) {
+          scopes_ = java.util.Collections.unmodifiableList(scopes_);
+        }
+        this.unknownFields = unknownFields.build();
+        makeExtensionsImmutable();
+      }
+    }
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALKey_descriptor;
     }
-    
+
     protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
         internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALKey_fieldAccessorTable;
+      return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALKey_fieldAccessorTable
+          .ensureFieldAccessorsInitialized(
+              org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey.class, org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey.Builder.class);
+    }
+
+    public static com.google.protobuf.Parser<WALKey> PARSER =
+        new com.google.protobuf.AbstractParser<WALKey>() {
+      public WALKey parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        return new WALKey(input, extensionRegistry);
+      }
+    };
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<WALKey> getParserForType() {
+      return PARSER;
     }
-    
+
     private int bitField0_;
     // required bytes encoded_region_name = 1;
     public static final int ENCODED_REGION_NAME_FIELD_NUMBER = 1;
     private com.google.protobuf.ByteString encodedRegionName_;
+    /**
+     * <code>required bytes encoded_region_name = 1;</code>
+     */
     public boolean hasEncodedRegionName() {
       return ((bitField0_ & 0x00000001) == 0x00000001);
     }
+    /**
+     * <code>required bytes encoded_region_name = 1;</code>
+     */
     public com.google.protobuf.ByteString getEncodedRegionName() {
       return encodedRegionName_;
     }
-    
+
     // required bytes table_name = 2;
     public static final int TABLE_NAME_FIELD_NUMBER = 2;
     private com.google.protobuf.ByteString tableName_;
+    /**
+     * <code>required bytes table_name = 2;</code>
+     */
     public boolean hasTableName() {
       return ((bitField0_ & 0x00000002) == 0x00000002);
     }
+    /**
+     * <code>required bytes table_name = 2;</code>
+     */
     public com.google.protobuf.ByteString getTableName() {
       return tableName_;
     }
-    
+
     // required uint64 log_sequence_number = 3;
     public static final int LOG_SEQUENCE_NUMBER_FIELD_NUMBER = 3;
     private long logSequenceNumber_;
+    /**
+     * <code>required uint64 log_sequence_number = 3;</code>
+     */
     public boolean hasLogSequenceNumber() {
       return ((bitField0_ & 0x00000004) == 0x00000004);
     }
+    /**
+     * <code>required uint64 log_sequence_number = 3;</code>
+     */
     public long getLogSequenceNumber() {
       return logSequenceNumber_;
     }
-    
+
     // required uint64 write_time = 4;
     public static final int WRITE_TIME_FIELD_NUMBER = 4;
     private long writeTime_;
+    /**
+     * <code>required uint64 write_time = 4;</code>
+     */
     public boolean hasWriteTime() {
       return ((bitField0_ & 0x00000008) == 0x00000008);
     }
+    /**
+     * <code>required uint64 write_time = 4;</code>
+     */
     public long getWriteTime() {
       return writeTime_;
     }
-    
+
     // optional .UUID cluster_id = 5;
     public static final int CLUSTER_ID_FIELD_NUMBER = 5;
     private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_;
+    /**
+     * <code>optional .UUID cluster_id = 5;</code>
+     */
     public boolean hasClusterId() {
       return ((bitField0_ & 0x00000010) == 0x00000010);
     }
+    /**
+     * <code>optional .UUID cluster_id = 5;</code>
+     */
     public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId() {
       return clusterId_;
     }
+    /**
+     * <code>optional .UUID cluster_id = 5;</code>
+     */
     public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder() {
       return clusterId_;
     }
-    
+
     // repeated .FamilyScope scopes = 6;
     public static final int SCOPES_FIELD_NUMBER = 6;
     private java.util.List<org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope> scopes_;
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     public java.util.List<org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope> getScopesList() {
       return scopes_;
     }
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder> 
         getScopesOrBuilderList() {
       return scopes_;
     }
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     public int getScopesCount() {
       return scopes_.size();
     }
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope getScopes(int index) {
       return scopes_.get(index);
     }
+    /**
+     * <code>repeated .FamilyScope scopes = 6;</code>
+     */
     public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder(
         int index) {
       return scopes_.get(index);
     }
-    
+
     // optional uint32 following_kv_count = 7;
     public static final int FOLLOWING_KV_COUNT_FIELD_NUMBER = 7;
     private int followingKvCount_;
+    /**
+     * <code>optional uint32 following_kv_count = 7;</code>
+     *
+     * <pre>
+     *
+     *optional CustomEntryType custom_entry_type = 8;
+     *
+     *enum CustomEntryType {
+     *COMPACTION = 0;
+     *}
+     * </pre>
+     */
     public boolean hasFollowingKvCount() {
       return ((bitField0_ & 0x00000020) == 0x00000020);
     }
+    /**
+     * <code>optional uint32 following_kv_count = 7;</code>
+     *
+     * <pre>
+     *
+     *optional CustomEntryType custom_entry_type = 8;
+     *
+     *enum CustomEntryType {
+     *COMPACTION = 0;
+     *}
+     * </pre>
+     */
     public int getFollowingKvCount() {
       return followingKvCount_;
     }
-    
+
     private void initFields() {
       encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
       tableName_ = com.google.protobuf.ByteString.EMPTY;
@@ -612,7 +947,7 @@ public final class WALProtos {
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
       if (isInitialized != -1) return isInitialized == 1;
-      
+
       if (!hasEncodedRegionName()) {
         memoizedIsInitialized = 0;
         return false;
@@ -644,7 +979,7 @@ public final class WALProtos {
       memoizedIsInitialized = 1;
       return true;
     }
-    
+
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
       getSerializedSize();
@@ -671,12 +1006,12 @@ public final class WALProtos {
       }
       getUnknownFields().writeTo(output);
     }
-    
+
     private int memoizedSerializedSize = -1;
     public int getSerializedSize() {
       int size = memoizedSerializedSize;
       if (size != -1) return size;
-    
+
       size = 0;
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
         size += com.google.protobuf.CodedOutputStream
@@ -710,14 +1045,14 @@ public final class WALProtos {
       memoizedSerializedSize = size;
       return size;
     }
-    
+
     private static final long serialVersionUID = 0L;
     @java.lang.Override
     protected java.lang.Object writeReplace()
         throws java.io.ObjectStreamException {
       return super.writeReplace();
     }
-    
+
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -727,7 +1062,7 @@ public final class WALProtos {
         return super.equals(obj);
       }
       org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey other = (org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey) obj;
-      
+
       boolean result = true;
       result = result && (hasEncodedRegionName() == other.hasEncodedRegionName());
       if (hasEncodedRegionName()) {
@@ -765,9 +1100,13 @@ public final class WALProtos {
           getUnknownFields().equals(other.getUnknownFields());
       return result;
     }
-    
+
+    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
+      if (memoizedHashCode != 0) {
+        return memoizedHashCode;
+      }
       int hash = 41;
       hash = (19 * hash) + getDescriptorForType().hashCode();
       if (hasEncodedRegionName()) {
@@ -799,89 +1138,83 @@ public final class WALProtos {
         hash = (53 * hash) + getFollowingKvCount();
       }
       hash = (29 * hash) + getUnknownFields().hashCode();
+      memoizedHashCode = hash;
       return hash;
     }
-    
+
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+      return PARSER.parseFrom(data);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseFrom(
         com.google.protobuf.ByteString data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(data, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseFrom(byte[] data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+      return PARSER.parseFrom(data);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseFrom(
         byte[] data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(data, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+      return PARSER.parseFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseDelimitedFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseDelimitedFrom(input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+      return PARSER.parseFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(input, extensionRegistry);
     }
-    
+
     public static Builder newBuilder() { return Builder.create(); }
     public Builder newBuilderForType() { return newBuilder(); }
     public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey prototype) {
       return newBuilder().mergeFrom(prototype);
     }
     public Builder toBuilder() { return newBuilder(this); }
-    
+
     @java.lang.Override
     protected Builder newBuilderForType(
         com.google.protobuf.GeneratedMessage.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
+    /**
+     * Protobuf type {@code WALKey}
+     *
+     * <pre>
+     * Protocol buffer version of HLogKey; see HLogKey comment, not really a key but WALEdit header for some KVs
+     * </pre>
+     */
     public static final class Builder extends
         com.google.protobuf.GeneratedMessage.Builder<Builder>
        implements org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKeyOrBuilder {
@@ -889,18 +1222,21 @@ public final class WALProtos {
           getDescriptor() {
         return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALKey_descriptor;
       }
-      
+
       protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
           internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALKey_fieldAccessorTable;
+        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALKey_fieldAccessorTable
+            .ensureFieldAccessorsInitialized(
+                org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey.class, org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey.Builder.class);
       }
-      
+
       // Construct using org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey.newBuilder()
       private Builder() {
         maybeForceBuilderInitialization();
       }
-      
-      private Builder(BuilderParent parent) {
+
+      private Builder(
+          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
@@ -913,7 +1249,7 @@ public final class WALProtos {
       private static Builder create() {
         return new Builder();
       }
-      
+
       public Builder clear() {
         super.clear();
         encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
@@ -940,20 +1276,20 @@ public final class WALProtos {
         bitField0_ = (bitField0_ & ~0x00000040);
         return this;
       }
-      
+
       public Builder clone() {
         return create().mergeFrom(buildPartial());
       }
-      
+
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey.getDescriptor();
+        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_WALKey_descriptor;
       }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey getDefaultInstanceForType() {
         return org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey.getDefaultInstance();
       }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey build() {
         org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey result = buildPartial();
         if (!result.isInitialized()) {
@@ -961,17 +1297,7 @@ public final class WALProtos {
         }
         return result;
       }
-      
-      private org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey buildParsed()
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(
-            result).asInvalidProtocolBufferException();
-        }
-        return result;
-      }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey buildPartial() {
         org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey result = new org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey(this);
         int from_bitField0_ = bitField0_;
@@ -1017,7 +1343,7 @@ public final class WALProtos {
         onBuilt();
         return result;
       }
-      
+
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey) {
           return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey)other);
@@ -1026,7 +1352,7 @@ public final class WALProtos {
           return this;
         }
       }
-      
+
       public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey other) {
         if (other == org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey.getDefaultInstance()) return this;
         if (other.hasEncodedRegionName()) {
@@ -1076,7 +1402,7 @@ public final class WALProtos {
         this.mergeUnknownFields(other.getUnknownFields());
         return this;
       }
-      
+
       public final boolean isInitialized() {
         if (!hasEncodedRegionName()) {
           
@@ -1108,84 +1434,43 @@ public final class WALProtos {
         }
         return true;
       }
-      
+
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder(
-            this.getUnknownFields());
-        while (true) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              this.setUnknownFields(unknownFields.build());
-              onChanged();
-              return this;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                this.setUnknownFields(unknownFields.build());
-                onChanged();
-                return this;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              encodedRegionName_ = input.readBytes();
-              break;
-            }
-            case 18: {
-              bitField0_ |= 0x00000002;
-              tableName_ = input.readBytes();
-              break;
-            }
-            case 24: {
-              bitField0_ |= 0x00000004;
-              logSequenceNumber_ = input.readUInt64();
-              break;
-            }
-            case 32: {
-              bitField0_ |= 0x00000008;
-              writeTime_ = input.readUInt64();
-              break;
-            }
-            case 42: {
-              org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.newBuilder();
-              if (hasClusterId()) {
-                subBuilder.mergeFrom(getClusterId());
-              }
-              input.readMessage(subBuilder, extensionRegistry);
-              setClusterId(subBuilder.buildPartial());
-              break;
-            }
-            case 50: {
-              org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder subBuilder = org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.newBuilder();
-              input.readMessage(subBuilder, extensionRegistry);
-              addScopes(subBuilder.buildPartial());
-              break;
-            }
-            case 56: {
-              bitField0_ |= 0x00000040;
-              followingKvCount_ = input.readUInt32();
-              break;
-            }
+        org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey parsedMessage = null;
+        try {
+          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.WALProtos.WALKey) e.getUnfinishedMessage();
+          throw e;
+        } finally {
+          if (parsedMessage != null) {
+            mergeFrom(parsedMessage);
           }
         }
+        return this;
       }
-      
       private int bitField0_;
-      
+
       // required bytes encoded_region_name = 1;
       private com.google.protobuf.ByteString encodedRegionName_ = com.google.protobuf.ByteString.EMPTY;
+      /**
+       * <code>required bytes encoded_region_name = 1;</code>
+       */
       public boolean hasEncodedRegionName() {
         return ((bitField0_ & 0x00000001) == 0x00000001);
       }
+      /**
+       * <code>required bytes encoded_region_name = 1;</code>
+       */
       public com.google.protobuf.ByteString getEncodedRegionName() {
         return encodedRegionName_;
       }
+      /**
+       * <code>required bytes encoded_region_name = 1;</code>
+       */
       public Builder setEncodedRegionName(com.google.protobuf.ByteString value) {
         if (value == null) {
     throw new NullPointerException();
@@ -1195,21 +1480,33 @@ public final class WALProtos {
         onChanged();
         return this;
       }
+      /**
+       * <code>required bytes encoded_region_name = 1;</code>
+       */
       public Builder clearEncodedRegionName() {
         bitField0_ = (bitField0_ & ~0x00000001);
         encodedRegionName_ = getDefaultInstance().getEncodedRegionName();
         onChanged();
         return this;
       }
-      
+
       // required bytes table_name = 2;
       private com.google.protobuf.ByteString tableName_ = com.google.protobuf.ByteString.EMPTY;
+      /**
+       * <code>required bytes table_name = 2;</code>
+       */
       public boolean hasTableName() {
         return ((bitField0_ & 0x00000002) == 0x00000002);
       }
+      /**
+       * <code>required bytes table_name = 2;</code>
+       */
       public com.google.protobuf.ByteString getTableName() {
         return tableName_;
       }
+      /**
+       * <code>required bytes table_name = 2;</code>
+       */
       public Builder setTableName(com.google.protobuf.ByteString value) {
         if (value == null) {
     throw new NullPointerException();
@@ -1219,62 +1516,95 @@ public final class WALProtos {
         onChanged();
         return this;
       }
+      /**
+       * <code>required bytes table_name = 2;</code>
+       */
       public Builder clearTableName() {
         bitField0_ = (bitField0_ & ~0x00000002);
         tableName_ = getDefaultInstance().getTableName();
         onChanged();
         return this;
       }
-      
+
       // required uint64 log_sequence_number = 3;
       private long logSequenceNumber_ ;
+      /**
+       * <code>required uint64 log_sequence_number = 3;</code>
+       */
       public boolean hasLogSequenceNumber() {
         return ((bitField0_ & 0x00000004) == 0x00000004);
       }
+      /**
+       * <code>required uint64 log_sequence_number = 3;</code>
+       */
       public long getLogSequenceNumber() {
         return logSequenceNumber_;
       }
+      /**
+       * <code>required uint64 log_sequence_number = 3;</code>
+       */
       public Builder setLogSequenceNumber(long value) {
         bitField0_ |= 0x00000004;
         logSequenceNumber_ = value;
         onChanged();
         return this;
       }
+      /**
+       * <code>required uint64 log_sequence_number = 3;</code>
+       */
       public Builder clearLogSequenceNumber() {
         bitField0_ = (bitField0_ & ~0x00000004);
         logSequenceNumber_ = 0L;
         onChanged();
         return this;
       }
-      
+
       // required uint64 write_time = 4;
       private long writeTime_ ;
+      /**
+       * <code>required uint64 write_time = 4;</code>
+       */
       public boolean hasWriteTime() {
         return ((bitField0_ & 0x00000008) == 0x00000008);
       }
+      /**
+       * <code>required uint64 write_time = 4;</code>
+       */
       public long getWriteTime() {
         return writeTime_;
       }
+      /**
+       * <code>required uint64 write_time = 4;</code>
+       */
       public Builder setWriteTime(long value) {
         bitField0_ |= 0x00000008;
         writeTime_ = value;
         onChanged();
         return this;
       }
+      /**
+       * <code>required uint64 write_time = 4;</code>
+       */
       public Builder clearWriteTime() {
         bitField0_ = (bitField0_ & ~0x00000008);
         writeTime_ = 0L;
         onChanged();
         return this;
       }
-      
+
       // optional .UUID cluster_id = 5;
       private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance();
       private com.google.protobuf.SingleFieldBuilder<
           org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> clusterIdBuilder_;
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       public boolean hasClusterId() {
         return ((bitField0_ & 0x00000010) == 0x00000010);
       }
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID getClusterId() {
         if (clusterIdBuilder_ == null) {
           return clusterId_;
@@ -1282,6 +1612,9 @@ public final class WALProtos {
           return clusterIdBuilder_.getMessage();
         }
       }
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       public Builder setClusterId(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID value) {
         if (clusterIdBuilder_ == null) {
           if (value == null) {
@@ -1295,6 +1628,9 @@ public final class WALProtos {
         bitField0_ |= 0x00000010;
         return this;
       }
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       public Builder setClusterId(
           org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder builderForValue) {
         if (clusterIdBuilder_ == null) {
@@ -1306,6 +1642,9 @@ public final class WALProtos {
         bitField0_ |= 0x00000010;
         return this;
       }
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       public Builder mergeClusterId(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID value) {
         if (clusterIdBuilder_ == null) {
           if (((bitField0_ & 0x00000010) == 0x00000010) &&
@@ -1322,6 +1661,9 @@ public final class WALProtos {
         bitField0_ |= 0x00000010;
         return this;
       }
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       public Builder clearClusterId() {
         if (clusterIdBuilder_ == null) {
           clusterId_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.getDefaultInstance();
@@ -1332,11 +1674,17 @@ public final class WALProtos {
         bitField0_ = (bitField0_ & ~0x00000010);
         return this;
       }
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder getClusterIdBuilder() {
         bitField0_ |= 0x00000010;
         onChanged();
         return getClusterIdFieldBuilder().getBuilder();
       }
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder getClusterIdOrBuilder() {
         if (clusterIdBuilder_ != null) {
           return clusterIdBuilder_.getMessageOrBuilder();
@@ -1344,6 +1692,9 @@ public final class WALProtos {
           return clusterId_;
         }
       }
+      /**
+       * <code>optional .UUID cluster_id = 5;</code>
+       */
       private com.google.protobuf.SingleFieldBuilder<
           org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUID.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.UUIDOrBuilder> 
           getClusterIdFieldBuilder() {
@@ -1357,7 +1708,7 @@ public final class WALProtos {
         }
         return clusterIdBuilder_;
       }
-      
+
       // repeated .FamilyScope scopes = 6;
       private java.util.List<org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope> scopes_ =
         java.util.Collections.emptyList();
@@ -1367,10 +1718,13 @@ public final class WALProtos {
           bitField0_ |= 0x00000020;
          }
       }
-      
+
       private com.google.protobuf.RepeatedFieldBuilder<
           org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder> scopesBuilder_;
-      
+
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public java.util.List<org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope> getScopesList() {
         if (scopesBuilder_ == null) {
           return java.util.Collections.unmodifiableList(scopes_);
@@ -1378,6 +1732,9 @@ public final class WALProtos {
           return scopesBuilder_.getMessageList();
         }
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public int getScopesCount() {
         if (scopesBuilder_ == null) {
           return scopes_.size();
@@ -1385,6 +1742,9 @@ public final class WALProtos {
           return scopesBuilder_.getCount();
         }
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope getScopes(int index) {
         if (scopesBuilder_ == null) {
           return scopes_.get(index);
@@ -1392,6 +1752,9 @@ public final class WALProtos {
           return scopesBuilder_.getMessage(index);
         }
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder setScopes(
           int index, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope value) {
         if (scopesBuilder_ == null) {
@@ -1406,6 +1769,9 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder setScopes(
           int index, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder builderForValue) {
         if (scopesBuilder_ == null) {
@@ -1417,6 +1783,9 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder addScopes(org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope value) {
         if (scopesBuilder_ == null) {
           if (value == null) {
@@ -1430,6 +1799,9 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder addScopes(
           int index, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope value) {
         if (scopesBuilder_ == null) {
@@ -1444,6 +1816,9 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder addScopes(
           org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder builderForValue) {
         if (scopesBuilder_ == null) {
@@ -1455,6 +1830,9 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder addScopes(
           int index, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder builderForValue) {
         if (scopesBuilder_ == null) {
@@ -1466,6 +1844,9 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder addAllScopes(
           java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope> values) {
         if (scopesBuilder_ == null) {
@@ -1477,6 +1858,9 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder clearScopes() {
         if (scopesBuilder_ == null) {
           scopes_ = java.util.Collections.emptyList();
@@ -1487,6 +1871,9 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public Builder removeScopes(int index) {
         if (scopesBuilder_ == null) {
           ensureScopesIsMutable();
@@ -1497,10 +1884,16 @@ public final class WALProtos {
         }
         return this;
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder getScopesBuilder(
           int index) {
         return getScopesFieldBuilder().getBuilder(index);
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder getScopesOrBuilder(
           int index) {
         if (scopesBuilder_ == null) {
@@ -1508,6 +1901,9 @@ public final class WALProtos {
           return scopesBuilder_.getMessageOrBuilder(index);
         }
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder> 
            getScopesOrBuilderList() {
         if (scopesBuilder_ != null) {
@@ -1516,15 +1912,24 @@ public final class WALProtos {
           return java.util.Collections.unmodifiableList(scopes_);
         }
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder addScopesBuilder() {
         return getScopesFieldBuilder().addBuilder(
             org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.getDefaultInstance());
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder addScopesBuilder(
           int index) {
         return getScopesFieldBuilder().addBuilder(
             index, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.getDefaultInstance());
       }
+      /**
+       * <code>repeated .FamilyScope scopes = 6;</code>
+       */
       public java.util.List<org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder> 
            getScopesBuilderList() {
         return getScopesFieldBuilder().getBuilderList();
@@ -1543,99 +1948,249 @@ public final class WALProtos {
         }
         return scopesBuilder_;
       }
-      
+
       // optional uint32 following_kv_count = 7;
       private int followingKvCount_ ;
+      /**
+       * <code>optional uint32 following_kv_count = 7;</code>
+       *
+       * <pre>
+       *
+       *optional CustomEntryType custom_entry_type = 8;
+       *
+       *enum CustomEntryType {
+       *COMPACTION = 0;
+       *}
+       * </pre>
+       */
       public boolean hasFollowingKvCount() {
         return ((bitField0_ & 0x00000040) == 0x00000040);
       }
+      /**
+       * <code>optional uint32 following_kv_count = 7;</code>
+       *
+       * <pre>
+       *
+       *optional CustomEntryType custom_entry_type = 8;
+       *
+       *enum CustomEntryType {
+       *COMPACTION = 0;
+       *}
+       * </pre>
+       */
       public int getFollowingKvCount() {
         return followingKvCount_;
       }
+      /**
+       * <code>optional uint32 following_kv_count = 7;</code>
+       *
+       * <pre>
+       *
+       *optional CustomEntryType custom_entry_type = 8;
+       *
+       *enum CustomEntryType {
+       *COMPACTION = 0;
+       *}
+       * </pre>
+       */
       public Builder setFollowingKvCount(int value) {
         bitField0_ |= 0x00000040;
         followingKvCount_ = value;
         onChanged();
         return this;
       }
+      /**
+       * <code>optional uint32 following_kv_count = 7;</code>
+       *
+       * <pre>
+       *
+       *optional CustomEntryType custom_entry_type = 8;
+       *
+       *enum CustomEntryType {
+       *COMPACTION = 0;
+       *}
+       * </pre>
+       */
       public Builder clearFollowingKvCount() {
         bitField0_ = (bitField0_ & ~0x00000040);
         followingKvCount_ = 0;
         onChanged();
         return this;
       }
-      
+
       // @@protoc_insertion_point(builder_scope:WALKey)
     }
-    
+
     static {
       defaultInstance = new WALKey(true);
       defaultInstance.initFields();
     }
-    
+
     // @@protoc_insertion_point(class_scope:WALKey)
   }
-  
+
   public interface FamilyScopeOrBuilder
       extends com.google.protobuf.MessageOrBuilder {
-    
+
     // required bytes family = 1;
+    /**
+     * <code>required bytes family = 1;</code>
+     */
     boolean hasFamily();
+    /**
+     * <code>required bytes family = 1;</code>
+     */
     com.google.protobuf.ByteString getFamily();
-    
+
     // required .ScopeType scope_type = 2;
+    /**
+     * <code>required .ScopeType scope_type = 2;</code>
+     */
     boolean hasScopeType();
+    /**
+     * <code>required .ScopeType scope_type = 2;</code>
+     */
     org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType();
   }
+  /**
+   * Protobuf type {@code FamilyScope}
+   */
   public static final class FamilyScope extends
       com.google.protobuf.GeneratedMessage
       implements FamilyScopeOrBuilder {
     // Use FamilyScope.newBuilder() to construct.
-    private FamilyScope(Builder builder) {
+    private FamilyScope(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
       super(builder);
+      this.unknownFields = builder.getUnknownFields();
     }
-    private FamilyScope(boolean noInit) {}
-    
+    private FamilyScope(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
     private static final FamilyScope defaultInstance;
     public static FamilyScope getDefaultInstance() {
       return defaultInstance;
     }
-    
+
     public FamilyScope getDefaultInstanceForType() {
       return defaultInstance;
     }
-    
+
+    private final com.google.protobuf.UnknownFieldSet unknownFields;
+    @java.lang.Override
+    public final com.google.protobuf.UnknownFieldSet
+        getUnknownFields() {
+      return this.unknownFields;
+    }
+    private FamilyScope(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      initFields();
+      int mutable_bitField0_ = 0;
+      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+          com.google.protobuf.UnknownFieldSet.newBuilder();
+      try {
+        boolean done = false;
+        while (!done) {
+          int tag = input.readTag();
+          switch (tag) {
+            case 0:
+              done = true;
+              break;
+            default: {
+              if (!parseUnknownField(input, unknownFields,
+                                     extensionRegistry, tag)) {
+                done = true;
+              }
+              break;
+            }
+            case 10: {
+              bitField0_ |= 0x00000001;
+              family_ = input.readBytes();
+              break;
+            }
+            case 16: {
+              int rawValue = input.readEnum();
+              org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType value = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.valueOf(rawValue);
+              if (value == null) {
+                unknownFields.mergeVarintField(2, rawValue);
+              } else {
+                bitField0_ |= 0x00000002;
+                scopeType_ = value;
+              }
+              break;
+            }
+          }
+        }
+      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+        throw e.setUnfinishedMessage(this);
+      } catch (java.io.IOException e) {
+        throw new com.google.protobuf.InvalidProtocolBufferException(
+            e.getMessage()).setUnfinishedMessage(this);
+      } finally {
+        this.unknownFields = unknownFields.build();
+        makeExtensionsImmutable();
+      }
+    }
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
       return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_FamilyScope_descriptor;
     }
-    
+
     protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
         internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_FamilyScope_fieldAccessorTable;
+      return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_FamilyScope_fieldAccessorTable
+          .ensureFieldAccessorsInitialized(
+              org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.class, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder.class);
+    }
+
+    public static com.google.protobuf.Parser<FamilyScope> PARSER =
+        new com.google.protobuf.AbstractParser<FamilyScope>() {
+      public FamilyScope parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        return new FamilyScope(input, extensionRegistry);
+      }
+    };
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<FamilyScope> getParserForType() {
+      return PARSER;
     }
-    
+
     private int bitField0_;
     // required bytes family = 1;
     public static final int FAMILY_FIELD_NUMBER = 1;
     private com.google.protobuf.ByteString family_;
+    /**
+     * <code>required bytes family = 1;</code>
+     */
     public boolean hasFamily() {
       return ((bitField0_ & 0x00000001) == 0x00000001);
     }
+    /**
+     * <code>required bytes family = 1;</code>
+     */
     public com.google.protobuf.ByteString getFamily() {
       return family_;
     }
-    
+
     // required .ScopeType scope_type = 2;
     public static final int SCOPE_TYPE_FIELD_NUMBER = 2;
     private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_;
+    /**
+     * <code>required .ScopeType scope_type = 2;</code>
+     */
     public boolean hasScopeType() {
       return ((bitField0_ & 0x00000002) == 0x00000002);
     }
+    /**
+     * <code>required .ScopeType scope_type = 2;</code>
+     */
     public org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType() {
       return scopeType_;
     }
-    
+
     private void initFields() {
       family_ = com.google.protobuf.ByteString.EMPTY;
       scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL;
@@ -1644,7 +2199,7 @@ public final class WALProtos {
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
       if (isInitialized != -1) return isInitialized == 1;
-      
+
       if (!hasFamily()) {
         memoizedIsInitialized = 0;
         return false;
@@ -1656,7 +2211,7 @@ public final class WALProtos {
       memoizedIsInitialized = 1;
       return true;
     }
-    
+
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
       getSerializedSize();
@@ -1668,12 +2223,12 @@ public final class WALProtos {
       }
       getUnknownFields().writeTo(output);
     }
-    
+
     private int memoizedSerializedSize = -1;
     public int getSerializedSize() {
       int size = memoizedSerializedSize;
       if (size != -1) return size;
-    
+
       size = 0;
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
         size += com.google.protobuf.CodedOutputStream
@@ -1687,14 +2242,14 @@ public final class WALProtos {
       memoizedSerializedSize = size;
       return size;
     }
-    
+
     private static final long serialVersionUID = 0L;
     @java.lang.Override
     protected java.lang.Object writeReplace()
         throws java.io.ObjectStreamException {
       return super.writeReplace();
     }
-    
+
     @java.lang.Override
     public boolean equals(final java.lang.Object obj) {
       if (obj == this) {
@@ -1704,7 +2259,7 @@ public final class WALProtos {
         return super.equals(obj);
       }
       org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope other = (org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope) obj;
-      
+
       boolean result = true;
       result = result && (hasFamily() == other.hasFamily());
       if (hasFamily()) {
@@ -1720,9 +2275,13 @@ public final class WALProtos {
           getUnknownFields().equals(other.getUnknownFields());
       return result;
     }
-    
+
+    private int memoizedHashCode = 0;
     @java.lang.Override
     public int hashCode() {
+      if (memoizedHashCode != 0) {
+        return memoizedHashCode;
+      }
       int hash = 41;
       hash = (19 * hash) + getDescriptorForType().hashCode();
       if (hasFamily()) {
@@ -1734,89 +2293,79 @@ public final class WALProtos {
         hash = (53 * hash) + hashEnum(getScopeType());
       }
       hash = (29 * hash) + getUnknownFields().hashCode();
+      memoizedHashCode = hash;
       return hash;
     }
-    
+
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+      return PARSER.parseFrom(data);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseFrom(
         com.google.protobuf.ByteString data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(data, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseFrom(byte[] data)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data).buildParsed();
+      return PARSER.parseFrom(data);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseFrom(
         byte[] data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
-      return newBuilder().mergeFrom(data, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(data, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseFrom(java.io.InputStream input)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+      return PARSER.parseFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseDelimitedFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      Builder builder = newBuilder();
-      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
-        return builder.buildParsed();
-      } else {
-        return null;
-      }
+      return PARSER.parseDelimitedFrom(input, extensionRegistry);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input).buildParsed();
+      return PARSER.parseFrom(input);
     }
     public static org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
-      return newBuilder().mergeFrom(input, extensionRegistry)
-               .buildParsed();
+      return PARSER.parseFrom(input, extensionRegistry);
     }
-    
+
     public static Builder newBuilder() { return Builder.create(); }
     public Builder newBuilderForType() { return newBuilder(); }
     public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope prototype) {
       return newBuilder().mergeFrom(prototype);
     }
     public Builder toBuilder() { return newBuilder(this); }
-    
+
     @java.lang.Override
     protected Builder newBuilderForType(
         com.google.protobuf.GeneratedMessage.BuilderParent parent) {
       Builder builder = new Builder(parent);
       return builder;
     }
+    /**
+     * Protobuf type {@code FamilyScope}
+     */
     public static final class Builder extends
         com.google.protobuf.GeneratedMessage.Builder<Builder>
        implements org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScopeOrBuilder {
@@ -1824,18 +2373,21 @@ public final class WALProtos {
           getDescriptor() {
         return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_FamilyScope_descriptor;
       }
-      
+
       protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
           internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_FamilyScope_fieldAccessorTable;
+        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_FamilyScope_fieldAccessorTable
+            .ensureFieldAccessorsInitialized(
+                org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.class, org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.Builder.class);
       }
-      
+
       // Construct using org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.newBuilder()
       private Builder() {
         maybeForceBuilderInitialization();
       }
-      
-      private Builder(BuilderParent parent) {
+
+      private Builder(
+          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
         super(parent);
         maybeForceBuilderInitialization();
       }
@@ -1846,7 +2398,7 @@ public final class WALProtos {
       private static Builder create() {
         return new Builder();
       }
-      
+
       public Builder clear() {
         super.clear();
         family_ = com.google.protobuf.ByteString.EMPTY;
@@ -1855,20 +2407,20 @@ public final class WALProtos {
         bitField0_ = (bitField0_ & ~0x00000002);
         return this;
       }
-      
+
       public Builder clone() {
         return create().mergeFrom(buildPartial());
       }
-      
+
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.getDescriptor();
+        return org.apache.hadoop.hbase.protobuf.generated.WALProtos.internal_static_FamilyScope_descriptor;
       }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope getDefaultInstanceForType() {
         return org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.getDefaultInstance();
       }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope build() {
         org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope result = buildPartial();
         if (!result.isInitialized()) {
@@ -1876,17 +2428,7 @@ public final class WALProtos {
         }
         return result;
       }
-      
-      private org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope buildParsed()
-          throws com.google.protobuf.InvalidProtocolBufferException {
-        org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope result = buildPartial();
-        if (!result.isInitialized()) {
-          throw newUninitializedMessageException(
-            result).asInvalidProtocolBufferException();
-        }
-        return result;
-      }
-      
+
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope buildPartial() {
         org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope result = new org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope(this);
         int from_bitField0_ = bitField0_;
@@ -1903,7 +2445,7 @@ public final class WALProtos {
         onBuilt();
         return result;
       }
-      
+
       public Builder mergeFrom(com.google.protobuf.Message other) {
         if (other instanceof org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope) {
           return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope)other);
@@ -1912,7 +2454,7 @@ public final class WALProtos {
           return this;
         }
       }
-      
+
       public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope other) {
         if (other == org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope.getDefaultInstance()) return this;
         if (other.hasFamily()) {
@@ -1924,7 +2466,7 @@ public final class WALProtos {
         this.mergeUnknownFields(other.getUnknownFields());
         return this;
       }
-      
+
       public final boolean isInitialized() {
         if (!hasFamily()) {
           
@@ -1936,60 +2478,43 @@ public final class WALProtos {
         }
         return true;
       }
-      
+
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
-          com.google.protobuf.UnknownFieldSet.newBuilder(
-            this.getUnknownFields());
-        while (true) {
-          int tag = input.readTag();
-          switch (tag) {
-            case 0:
-              this.setUnknownFields(unknownFields.build());
-              onChanged();
-              return this;
-            default: {
-              if (!parseUnknownField(input, unknownFields,
-                                     extensionRegistry, tag)) {
-                this.setUnknownFields(unknownFields.build());
-                onChanged();
-                return this;
-              }
-              break;
-            }
-            case 10: {
-              bitField0_ |= 0x00000001;
-              family_ = input.readBytes();
-              break;
-            }
-            case 16: {
-              int rawValue = input.readEnum();
-              org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType value = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.valueOf(rawValue);
-              if (value == null) {
-                unknownFields.mergeVarintField(2, rawValue);
-              } else {
-                bitField0_ |= 0x00000002;
-                scopeType_ = value;
-              }
-              break;
-            }
+        org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope parsedMessage = null;
+        try {
+          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.WALProtos.FamilyScope) e.getUnfinishedMessage();
+          throw e;
+        } finally {
+          if (parsedMessage != null) {
+            mergeFrom(parsedMessage);
           }
         }
+        return this;
       }
-      
       private int bitField0_;
-      
+
       // required bytes family = 1;
       private com.google.protobuf.ByteString family_ = com.google.protobuf.ByteString.EMPTY;
+      /**
+       * <code>required bytes family = 1;</code>
+       */
       public boolean hasFamily() {
         return ((bitField0_ & 0x00000001) == 0x00000001);
       }
+      /**
+       * <code>required bytes family = 1;</code>
+       */
       public com.google.protobuf.ByteString getFamily() {
         return family_;
       }
+      /**
+       * <code>required bytes family = 1;</code>
+       */
       public Builder setFamily(com.google.protobuf.ByteString value) {
         if (value == null) {
     throw new NullPointerException();
@@ -1999,21 +2524,33 @@ public final class WALProtos {
         onChanged();
         return this;
       }
+      /**
+       * <code>required bytes family = 1;</code>
+       */
       public Builder clearFamily() {
         bitField0_ = (bitField0_ & ~0x00000001);
         family_ = getDefaultInstance().getFamily();
         onChanged();
         return this;
       }
-      
+
       // required .ScopeType scope_type = 2;
       private org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType scopeType_ = org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType.REPLICATION_SCOPE_LOCAL;
+      /**
+       * <code>required .ScopeType scope_type = 2;</code>
+       */
       public boolean hasScopeType() {
         return ((bitField0_ & 0x00000002) == 0x00000002);
       }
+      /**
+       * <code>required .ScopeType scope_type = 2;</code>
+       */
       public org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType getScopeType() {
         return scopeType_;
       }
+      /**
+       * <code>required .ScopeType scope_type = 2;</code>
+       */
       public Builder setScopeType(org.apache.hadoop.hbase.protobuf.generated.WALProtos.ScopeType value) {
         if (value == null) {
           throw new NullPointerException();
@@ -2023,172 +2560,423 @@ public final class WALProtos {
         onChanged();
         return this;
       }
+      /**

[... 1758 lines stripped ...]