You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by to...@apache.org on 2011/08/24 00:34:23 UTC

svn commit: r1160912 [3/6] - in /hbase/trunk: ./ src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/

Modified: hbase/trunk/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java
URL: http://svn.apache.org/viewvc/hbase/trunk/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java?rev=1160912&r1=1160911&r2=1160912&view=diff
==============================================================================
--- hbase/trunk/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java (original)
+++ hbase/trunk/src/main/java/org/apache/hadoop/hbase/rest/protobuf/generated/StorageClusterStatusMessage.java Tue Aug 23 22:34:23 2011
@@ -8,11 +8,42 @@ public final class StorageClusterStatusM
   public static void registerAllExtensions(
       com.google.protobuf.ExtensionRegistry registry) {
   }
+  public interface StorageClusterStatusOrBuilder
+      extends com.google.protobuf.MessageOrBuilder {
+    
+    // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1;
+    java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> 
+        getLiveNodesList();
+    org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index);
+    int getLiveNodesCount();
+    java.util.List<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder> 
+        getLiveNodesOrBuilderList();
+    org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder getLiveNodesOrBuilder(
+        int index);
+    
+    // repeated string deadNodes = 2;
+    java.util.List<String> getDeadNodesList();
+    int getDeadNodesCount();
+    String getDeadNodes(int index);
+    
+    // optional int32 regions = 3;
+    boolean hasRegions();
+    int getRegions();
+    
+    // optional int32 requests = 4;
+    boolean hasRequests();
+    int getRequests();
+    
+    // optional double averageLoad = 5;
+    boolean hasAverageLoad();
+    double getAverageLoad();
+  }
   public static final class StorageClusterStatus extends
-      com.google.protobuf.GeneratedMessage {
+      com.google.protobuf.GeneratedMessage
+      implements StorageClusterStatusOrBuilder {
     // Use StorageClusterStatus.newBuilder() to construct.
-    private StorageClusterStatus() {
-      initFields();
+    private StorageClusterStatus(Builder builder) {
+      super(builder);
     }
     private StorageClusterStatus(boolean noInit) {}
     
@@ -35,11 +66,39 @@ public final class StorageClusterStatusM
       return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable;
     }
     
+    public interface RegionOrBuilder
+        extends com.google.protobuf.MessageOrBuilder {
+      
+      // required bytes name = 1;
+      boolean hasName();
+      com.google.protobuf.ByteString getName();
+      
+      // optional int32 stores = 2;
+      boolean hasStores();
+      int getStores();
+      
+      // optional int32 storefiles = 3;
+      boolean hasStorefiles();
+      int getStorefiles();
+      
+      // optional int32 storefileSizeMB = 4;
+      boolean hasStorefileSizeMB();
+      int getStorefileSizeMB();
+      
+      // optional int32 memstoreSizeMB = 5;
+      boolean hasMemstoreSizeMB();
+      int getMemstoreSizeMB();
+      
+      // optional int32 storefileIndexSizeMB = 6;
+      boolean hasStorefileIndexSizeMB();
+      int getStorefileIndexSizeMB();
+    }
     public static final class Region extends
-        com.google.protobuf.GeneratedMessage {
+        com.google.protobuf.GeneratedMessage
+        implements RegionOrBuilder {
       // Use Region.newBuilder() to construct.
-      private Region() {
-        initFields();
+      private Region(Builder builder) {
+        super(builder);
       }
       private Region(boolean noInit) {}
       
@@ -62,75 +121,108 @@ public final class StorageClusterStatusM
         return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable;
       }
       
+      private int bitField0_;
       // required bytes name = 1;
       public static final int NAME_FIELD_NUMBER = 1;
-      private boolean hasName;
-      private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY;
-      public boolean hasName() { return hasName; }
-      public com.google.protobuf.ByteString getName() { return name_; }
+      private com.google.protobuf.ByteString name_;
+      public boolean hasName() {
+        return ((bitField0_ & 0x00000001) == 0x00000001);
+      }
+      public com.google.protobuf.ByteString getName() {
+        return name_;
+      }
       
       // optional int32 stores = 2;
       public static final int STORES_FIELD_NUMBER = 2;
-      private boolean hasStores;
-      private int stores_ = 0;
-      public boolean hasStores() { return hasStores; }
-      public int getStores() { return stores_; }
+      private int stores_;
+      public boolean hasStores() {
+        return ((bitField0_ & 0x00000002) == 0x00000002);
+      }
+      public int getStores() {
+        return stores_;
+      }
       
       // optional int32 storefiles = 3;
       public static final int STOREFILES_FIELD_NUMBER = 3;
-      private boolean hasStorefiles;
-      private int storefiles_ = 0;
-      public boolean hasStorefiles() { return hasStorefiles; }
-      public int getStorefiles() { return storefiles_; }
+      private int storefiles_;
+      public boolean hasStorefiles() {
+        return ((bitField0_ & 0x00000004) == 0x00000004);
+      }
+      public int getStorefiles() {
+        return storefiles_;
+      }
       
       // optional int32 storefileSizeMB = 4;
       public static final int STOREFILESIZEMB_FIELD_NUMBER = 4;
-      private boolean hasStorefileSizeMB;
-      private int storefileSizeMB_ = 0;
-      public boolean hasStorefileSizeMB() { return hasStorefileSizeMB; }
-      public int getStorefileSizeMB() { return storefileSizeMB_; }
+      private int storefileSizeMB_;
+      public boolean hasStorefileSizeMB() {
+        return ((bitField0_ & 0x00000008) == 0x00000008);
+      }
+      public int getStorefileSizeMB() {
+        return storefileSizeMB_;
+      }
       
       // optional int32 memstoreSizeMB = 5;
       public static final int MEMSTORESIZEMB_FIELD_NUMBER = 5;
-      private boolean hasMemstoreSizeMB;
-      private int memstoreSizeMB_ = 0;
-      public boolean hasMemstoreSizeMB() { return hasMemstoreSizeMB; }
-      public int getMemstoreSizeMB() { return memstoreSizeMB_; }
+      private int memstoreSizeMB_;
+      public boolean hasMemstoreSizeMB() {
+        return ((bitField0_ & 0x00000010) == 0x00000010);
+      }
+      public int getMemstoreSizeMB() {
+        return memstoreSizeMB_;
+      }
       
       // optional int32 storefileIndexSizeMB = 6;
       public static final int STOREFILEINDEXSIZEMB_FIELD_NUMBER = 6;
-      private boolean hasStorefileIndexSizeMB;
-      private int storefileIndexSizeMB_ = 0;
-      public boolean hasStorefileIndexSizeMB() { return hasStorefileIndexSizeMB; }
-      public int getStorefileIndexSizeMB() { return storefileIndexSizeMB_; }
+      private int storefileIndexSizeMB_;
+      public boolean hasStorefileIndexSizeMB() {
+        return ((bitField0_ & 0x00000020) == 0x00000020);
+      }
+      public int getStorefileIndexSizeMB() {
+        return storefileIndexSizeMB_;
+      }
       
       private void initFields() {
+        name_ = com.google.protobuf.ByteString.EMPTY;
+        stores_ = 0;
+        storefiles_ = 0;
+        storefileSizeMB_ = 0;
+        memstoreSizeMB_ = 0;
+        storefileIndexSizeMB_ = 0;
       }
+      private byte memoizedIsInitialized = -1;
       public final boolean isInitialized() {
-        if (!hasName) return false;
+        byte isInitialized = memoizedIsInitialized;
+        if (isInitialized != -1) return isInitialized == 1;
+        
+        if (!hasName()) {
+          memoizedIsInitialized = 0;
+          return false;
+        }
+        memoizedIsInitialized = 1;
         return true;
       }
       
       public void writeTo(com.google.protobuf.CodedOutputStream output)
                           throws java.io.IOException {
         getSerializedSize();
-        if (hasName()) {
-          output.writeBytes(1, getName());
+        if (((bitField0_ & 0x00000001) == 0x00000001)) {
+          output.writeBytes(1, name_);
         }
-        if (hasStores()) {
-          output.writeInt32(2, getStores());
+        if (((bitField0_ & 0x00000002) == 0x00000002)) {
+          output.writeInt32(2, stores_);
         }
-        if (hasStorefiles()) {
-          output.writeInt32(3, getStorefiles());
+        if (((bitField0_ & 0x00000004) == 0x00000004)) {
+          output.writeInt32(3, storefiles_);
         }
-        if (hasStorefileSizeMB()) {
-          output.writeInt32(4, getStorefileSizeMB());
+        if (((bitField0_ & 0x00000008) == 0x00000008)) {
+          output.writeInt32(4, storefileSizeMB_);
         }
-        if (hasMemstoreSizeMB()) {
-          output.writeInt32(5, getMemstoreSizeMB());
+        if (((bitField0_ & 0x00000010) == 0x00000010)) {
+          output.writeInt32(5, memstoreSizeMB_);
         }
-        if (hasStorefileIndexSizeMB()) {
-          output.writeInt32(6, getStorefileIndexSizeMB());
+        if (((bitField0_ & 0x00000020) == 0x00000020)) {
+          output.writeInt32(6, storefileIndexSizeMB_);
         }
         getUnknownFields().writeTo(output);
       }
@@ -141,35 +233,42 @@ public final class StorageClusterStatusM
         if (size != -1) return size;
       
         size = 0;
-        if (hasName()) {
+        if (((bitField0_ & 0x00000001) == 0x00000001)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeBytesSize(1, getName());
+            .computeBytesSize(1, name_);
         }
-        if (hasStores()) {
+        if (((bitField0_ & 0x00000002) == 0x00000002)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(2, getStores());
+            .computeInt32Size(2, stores_);
         }
-        if (hasStorefiles()) {
+        if (((bitField0_ & 0x00000004) == 0x00000004)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(3, getStorefiles());
+            .computeInt32Size(3, storefiles_);
         }
-        if (hasStorefileSizeMB()) {
+        if (((bitField0_ & 0x00000008) == 0x00000008)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(4, getStorefileSizeMB());
+            .computeInt32Size(4, storefileSizeMB_);
         }
-        if (hasMemstoreSizeMB()) {
+        if (((bitField0_ & 0x00000010) == 0x00000010)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(5, getMemstoreSizeMB());
+            .computeInt32Size(5, memstoreSizeMB_);
         }
-        if (hasStorefileIndexSizeMB()) {
+        if (((bitField0_ & 0x00000020) == 0x00000020)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(6, getStorefileIndexSizeMB());
+            .computeInt32Size(6, storefileIndexSizeMB_);
         }
         size += getUnknownFields().getSerializedSize();
         memoizedSerializedSize = size;
         return size;
       }
       
+      private static final long serialVersionUID = 0L;
+      @java.lang.Override
+      protected java.lang.Object writeReplace()
+          throws java.io.ObjectStreamException {
+        return super.writeReplace();
+      }
+      
       public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region parseFrom(
           com.google.protobuf.ByteString data)
           throws com.google.protobuf.InvalidProtocolBufferException {
@@ -244,34 +343,61 @@ public final class StorageClusterStatusM
       }
       public Builder toBuilder() { return newBuilder(this); }
       
+      @java.lang.Override
+      protected Builder newBuilderForType(
+          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        Builder builder = new Builder(parent);
+        return builder;
+      }
       public static final class Builder extends
-          com.google.protobuf.GeneratedMessage.Builder<Builder> {
-        private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region result;
+          com.google.protobuf.GeneratedMessage.Builder<Builder>
+         implements org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder {
+        public static final com.google.protobuf.Descriptors.Descriptor
+            getDescriptor() {
+          return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_descriptor;
+        }
         
-        // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.newBuilder()
-        private Builder() {}
+        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+            internalGetFieldAccessorTable() {
+          return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Region_fieldAccessorTable;
+        }
         
-        private static Builder create() {
-          Builder builder = new Builder();
-          builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region();
-          return builder;
+        // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.newBuilder()
+        private Builder() {
+          maybeForceBuilderInitialization();
         }
         
-        protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region internalGetResult() {
-          return result;
+        private Builder(BuilderParent parent) {
+          super(parent);
+          maybeForceBuilderInitialization();
+        }
+        private void maybeForceBuilderInitialization() {
+          if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+          }
+        }
+        private static Builder create() {
+          return new Builder();
         }
         
         public Builder clear() {
-          if (result == null) {
-            throw new IllegalStateException(
-              "Cannot call clear() after build().");
-          }
-          result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region();
+          super.clear();
+          name_ = com.google.protobuf.ByteString.EMPTY;
+          bitField0_ = (bitField0_ & ~0x00000001);
+          stores_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000002);
+          storefiles_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000004);
+          storefileSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000008);
+          memstoreSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000010);
+          storefileIndexSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000020);
           return this;
         }
         
         public Builder clone() {
-          return create().mergeFrom(result);
+          return create().mergeFrom(buildPartial());
         }
         
         public com.google.protobuf.Descriptors.Descriptor
@@ -283,33 +409,55 @@ public final class StorageClusterStatusM
           return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance();
         }
         
-        public boolean isInitialized() {
-          return result.isInitialized();
-        }
         public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region build() {
-          if (result != null && !isInitialized()) {
+          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region result = buildPartial();
+          if (!result.isInitialized()) {
             throw newUninitializedMessageException(result);
           }
-          return buildPartial();
+          return result;
         }
         
         private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region buildParsed()
             throws com.google.protobuf.InvalidProtocolBufferException {
-          if (!isInitialized()) {
+          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region result = buildPartial();
+          if (!result.isInitialized()) {
             throw newUninitializedMessageException(
               result).asInvalidProtocolBufferException();
           }
-          return buildPartial();
+          return result;
         }
         
         public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region buildPartial() {
-          if (result == null) {
-            throw new IllegalStateException(
-              "build() has already been called on this Builder.");
-          }
-          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region returnMe = result;
-          result = null;
-          return returnMe;
+          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region(this);
+          int from_bitField0_ = bitField0_;
+          int to_bitField0_ = 0;
+          if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+            to_bitField0_ |= 0x00000001;
+          }
+          result.name_ = name_;
+          if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+            to_bitField0_ |= 0x00000002;
+          }
+          result.stores_ = stores_;
+          if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+            to_bitField0_ |= 0x00000004;
+          }
+          result.storefiles_ = storefiles_;
+          if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
+            to_bitField0_ |= 0x00000008;
+          }
+          result.storefileSizeMB_ = storefileSizeMB_;
+          if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
+            to_bitField0_ |= 0x00000010;
+          }
+          result.memstoreSizeMB_ = memstoreSizeMB_;
+          if (((from_bitField0_ & 0x00000020) == 0x00000020)) {
+            to_bitField0_ |= 0x00000020;
+          }
+          result.storefileIndexSizeMB_ = storefileIndexSizeMB_;
+          result.bitField0_ = to_bitField0_;
+          onBuilt();
+          return result;
         }
         
         public Builder mergeFrom(com.google.protobuf.Message other) {
@@ -345,6 +493,14 @@ public final class StorageClusterStatusM
           return this;
         }
         
+        public final boolean isInitialized() {
+          if (!hasName()) {
+            
+            return false;
+          }
+          return true;
+        }
+        
         public Builder mergeFrom(
             com.google.protobuf.CodedInputStream input,
             com.google.protobuf.ExtensionRegistryLite extensionRegistry)
@@ -357,152 +513,179 @@ public final class StorageClusterStatusM
             switch (tag) {
               case 0:
                 this.setUnknownFields(unknownFields.build());
+                onChanged();
                 return this;
               default: {
                 if (!parseUnknownField(input, unknownFields,
                                        extensionRegistry, tag)) {
                   this.setUnknownFields(unknownFields.build());
+                  onChanged();
                   return this;
                 }
                 break;
               }
               case 10: {
-                setName(input.readBytes());
+                bitField0_ |= 0x00000001;
+                name_ = input.readBytes();
                 break;
               }
               case 16: {
-                setStores(input.readInt32());
+                bitField0_ |= 0x00000002;
+                stores_ = input.readInt32();
                 break;
               }
               case 24: {
-                setStorefiles(input.readInt32());
+                bitField0_ |= 0x00000004;
+                storefiles_ = input.readInt32();
                 break;
               }
               case 32: {
-                setStorefileSizeMB(input.readInt32());
+                bitField0_ |= 0x00000008;
+                storefileSizeMB_ = input.readInt32();
                 break;
               }
               case 40: {
-                setMemstoreSizeMB(input.readInt32());
+                bitField0_ |= 0x00000010;
+                memstoreSizeMB_ = input.readInt32();
                 break;
               }
               case 48: {
-                setStorefileIndexSizeMB(input.readInt32());
+                bitField0_ |= 0x00000020;
+                storefileIndexSizeMB_ = input.readInt32();
                 break;
               }
             }
           }
         }
         
+        private int bitField0_;
         
         // required bytes name = 1;
+        private com.google.protobuf.ByteString name_ = com.google.protobuf.ByteString.EMPTY;
         public boolean hasName() {
-          return result.hasName();
+          return ((bitField0_ & 0x00000001) == 0x00000001);
         }
         public com.google.protobuf.ByteString getName() {
-          return result.getName();
+          return name_;
         }
         public Builder setName(com.google.protobuf.ByteString value) {
           if (value == null) {
     throw new NullPointerException();
   }
-  result.hasName = true;
-          result.name_ = value;
+  bitField0_ |= 0x00000001;
+          name_ = value;
+          onChanged();
           return this;
         }
         public Builder clearName() {
-          result.hasName = false;
-          result.name_ = getDefaultInstance().getName();
+          bitField0_ = (bitField0_ & ~0x00000001);
+          name_ = getDefaultInstance().getName();
+          onChanged();
           return this;
         }
         
         // optional int32 stores = 2;
+        private int stores_ ;
         public boolean hasStores() {
-          return result.hasStores();
+          return ((bitField0_ & 0x00000002) == 0x00000002);
         }
         public int getStores() {
-          return result.getStores();
+          return stores_;
         }
         public Builder setStores(int value) {
-          result.hasStores = true;
-          result.stores_ = value;
+          bitField0_ |= 0x00000002;
+          stores_ = value;
+          onChanged();
           return this;
         }
         public Builder clearStores() {
-          result.hasStores = false;
-          result.stores_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000002);
+          stores_ = 0;
+          onChanged();
           return this;
         }
         
         // optional int32 storefiles = 3;
+        private int storefiles_ ;
         public boolean hasStorefiles() {
-          return result.hasStorefiles();
+          return ((bitField0_ & 0x00000004) == 0x00000004);
         }
         public int getStorefiles() {
-          return result.getStorefiles();
+          return storefiles_;
         }
         public Builder setStorefiles(int value) {
-          result.hasStorefiles = true;
-          result.storefiles_ = value;
+          bitField0_ |= 0x00000004;
+          storefiles_ = value;
+          onChanged();
           return this;
         }
         public Builder clearStorefiles() {
-          result.hasStorefiles = false;
-          result.storefiles_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000004);
+          storefiles_ = 0;
+          onChanged();
           return this;
         }
         
         // optional int32 storefileSizeMB = 4;
+        private int storefileSizeMB_ ;
         public boolean hasStorefileSizeMB() {
-          return result.hasStorefileSizeMB();
+          return ((bitField0_ & 0x00000008) == 0x00000008);
         }
         public int getStorefileSizeMB() {
-          return result.getStorefileSizeMB();
+          return storefileSizeMB_;
         }
         public Builder setStorefileSizeMB(int value) {
-          result.hasStorefileSizeMB = true;
-          result.storefileSizeMB_ = value;
+          bitField0_ |= 0x00000008;
+          storefileSizeMB_ = value;
+          onChanged();
           return this;
         }
         public Builder clearStorefileSizeMB() {
-          result.hasStorefileSizeMB = false;
-          result.storefileSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000008);
+          storefileSizeMB_ = 0;
+          onChanged();
           return this;
         }
         
         // optional int32 memstoreSizeMB = 5;
+        private int memstoreSizeMB_ ;
         public boolean hasMemstoreSizeMB() {
-          return result.hasMemstoreSizeMB();
+          return ((bitField0_ & 0x00000010) == 0x00000010);
         }
         public int getMemstoreSizeMB() {
-          return result.getMemstoreSizeMB();
+          return memstoreSizeMB_;
         }
         public Builder setMemstoreSizeMB(int value) {
-          result.hasMemstoreSizeMB = true;
-          result.memstoreSizeMB_ = value;
+          bitField0_ |= 0x00000010;
+          memstoreSizeMB_ = value;
+          onChanged();
           return this;
         }
         public Builder clearMemstoreSizeMB() {
-          result.hasMemstoreSizeMB = false;
-          result.memstoreSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000010);
+          memstoreSizeMB_ = 0;
+          onChanged();
           return this;
         }
         
         // optional int32 storefileIndexSizeMB = 6;
+        private int storefileIndexSizeMB_ ;
         public boolean hasStorefileIndexSizeMB() {
-          return result.hasStorefileIndexSizeMB();
+          return ((bitField0_ & 0x00000020) == 0x00000020);
         }
         public int getStorefileIndexSizeMB() {
-          return result.getStorefileIndexSizeMB();
+          return storefileIndexSizeMB_;
         }
         public Builder setStorefileIndexSizeMB(int value) {
-          result.hasStorefileIndexSizeMB = true;
-          result.storefileIndexSizeMB_ = value;
+          bitField0_ |= 0x00000020;
+          storefileIndexSizeMB_ = value;
+          onChanged();
           return this;
         }
         public Builder clearStorefileIndexSizeMB() {
-          result.hasStorefileIndexSizeMB = false;
-          result.storefileIndexSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000020);
+          storefileIndexSizeMB_ = 0;
+          onChanged();
           return this;
         }
         
@@ -511,18 +694,51 @@ public final class StorageClusterStatusM
       
       static {
         defaultInstance = new Region(true);
-        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit();
         defaultInstance.initFields();
       }
       
       // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region)
     }
     
+    public interface NodeOrBuilder
+        extends com.google.protobuf.MessageOrBuilder {
+      
+      // required string name = 1;
+      boolean hasName();
+      String getName();
+      
+      // optional int64 startCode = 2;
+      boolean hasStartCode();
+      long getStartCode();
+      
+      // optional int32 requests = 3;
+      boolean hasRequests();
+      int getRequests();
+      
+      // optional int32 heapSizeMB = 4;
+      boolean hasHeapSizeMB();
+      int getHeapSizeMB();
+      
+      // optional int32 maxHeapSizeMB = 5;
+      boolean hasMaxHeapSizeMB();
+      int getMaxHeapSizeMB();
+      
+      // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6;
+      java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> 
+          getRegionsList();
+      org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index);
+      int getRegionsCount();
+      java.util.List<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder> 
+          getRegionsOrBuilderList();
+      org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder getRegionsOrBuilder(
+          int index);
+    }
     public static final class Node extends
-        com.google.protobuf.GeneratedMessage {
+        com.google.protobuf.GeneratedMessage
+        implements NodeOrBuilder {
       // Use Node.newBuilder() to construct.
-      private Node() {
-        initFields();
+      private Node(Builder builder) {
+        super(builder);
       }
       private Node(boolean noInit) {}
       
@@ -545,83 +761,147 @@ public final class StorageClusterStatusM
         return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable;
       }
       
+      private int bitField0_;
       // required string name = 1;
       public static final int NAME_FIELD_NUMBER = 1;
-      private boolean hasName;
-      private java.lang.String name_ = "";
-      public boolean hasName() { return hasName; }
-      public java.lang.String getName() { return name_; }
+      private java.lang.Object name_;
+      public boolean hasName() {
+        return ((bitField0_ & 0x00000001) == 0x00000001);
+      }
+      public String getName() {
+        java.lang.Object ref = name_;
+        if (ref instanceof String) {
+          return (String) ref;
+        } else {
+          com.google.protobuf.ByteString bs = 
+              (com.google.protobuf.ByteString) ref;
+          String s = bs.toStringUtf8();
+          if (com.google.protobuf.Internal.isValidUtf8(bs)) {
+            name_ = s;
+          }
+          return s;
+        }
+      }
+      private com.google.protobuf.ByteString getNameBytes() {
+        java.lang.Object ref = name_;
+        if (ref instanceof String) {
+          com.google.protobuf.ByteString b = 
+              com.google.protobuf.ByteString.copyFromUtf8((String) ref);
+          name_ = b;
+          return b;
+        } else {
+          return (com.google.protobuf.ByteString) ref;
+        }
+      }
       
       // optional int64 startCode = 2;
       public static final int STARTCODE_FIELD_NUMBER = 2;
-      private boolean hasStartCode;
-      private long startCode_ = 0L;
-      public boolean hasStartCode() { return hasStartCode; }
-      public long getStartCode() { return startCode_; }
+      private long startCode_;
+      public boolean hasStartCode() {
+        return ((bitField0_ & 0x00000002) == 0x00000002);
+      }
+      public long getStartCode() {
+        return startCode_;
+      }
       
       // optional int32 requests = 3;
       public static final int REQUESTS_FIELD_NUMBER = 3;
-      private boolean hasRequests;
-      private int requests_ = 0;
-      public boolean hasRequests() { return hasRequests; }
-      public int getRequests() { return requests_; }
+      private int requests_;
+      public boolean hasRequests() {
+        return ((bitField0_ & 0x00000004) == 0x00000004);
+      }
+      public int getRequests() {
+        return requests_;
+      }
       
       // optional int32 heapSizeMB = 4;
       public static final int HEAPSIZEMB_FIELD_NUMBER = 4;
-      private boolean hasHeapSizeMB;
-      private int heapSizeMB_ = 0;
-      public boolean hasHeapSizeMB() { return hasHeapSizeMB; }
-      public int getHeapSizeMB() { return heapSizeMB_; }
+      private int heapSizeMB_;
+      public boolean hasHeapSizeMB() {
+        return ((bitField0_ & 0x00000008) == 0x00000008);
+      }
+      public int getHeapSizeMB() {
+        return heapSizeMB_;
+      }
       
       // optional int32 maxHeapSizeMB = 5;
       public static final int MAXHEAPSIZEMB_FIELD_NUMBER = 5;
-      private boolean hasMaxHeapSizeMB;
-      private int maxHeapSizeMB_ = 0;
-      public boolean hasMaxHeapSizeMB() { return hasMaxHeapSizeMB; }
-      public int getMaxHeapSizeMB() { return maxHeapSizeMB_; }
+      private int maxHeapSizeMB_;
+      public boolean hasMaxHeapSizeMB() {
+        return ((bitField0_ & 0x00000010) == 0x00000010);
+      }
+      public int getMaxHeapSizeMB() {
+        return maxHeapSizeMB_;
+      }
       
       // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6;
       public static final int REGIONS_FIELD_NUMBER = 6;
-      private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> regions_ =
-        java.util.Collections.emptyList();
+      private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> regions_;
       public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> getRegionsList() {
         return regions_;
       }
-      public int getRegionsCount() { return regions_.size(); }
+      public java.util.List<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder> 
+          getRegionsOrBuilderList() {
+        return regions_;
+      }
+      public int getRegionsCount() {
+        return regions_.size();
+      }
       public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index) {
         return regions_.get(index);
       }
+      public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder getRegionsOrBuilder(
+          int index) {
+        return regions_.get(index);
+      }
       
       private void initFields() {
+        name_ = "";
+        startCode_ = 0L;
+        requests_ = 0;
+        heapSizeMB_ = 0;
+        maxHeapSizeMB_ = 0;
+        regions_ = java.util.Collections.emptyList();
       }
+      private byte memoizedIsInitialized = -1;
       public final boolean isInitialized() {
-        if (!hasName) return false;
-        for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
-          if (!element.isInitialized()) return false;
+        byte isInitialized = memoizedIsInitialized;
+        if (isInitialized != -1) return isInitialized == 1;
+        
+        if (!hasName()) {
+          memoizedIsInitialized = 0;
+          return false;
+        }
+        for (int i = 0; i < getRegionsCount(); i++) {
+          if (!getRegions(i).isInitialized()) {
+            memoizedIsInitialized = 0;
+            return false;
+          }
         }
+        memoizedIsInitialized = 1;
         return true;
       }
       
       public void writeTo(com.google.protobuf.CodedOutputStream output)
                           throws java.io.IOException {
         getSerializedSize();
-        if (hasName()) {
-          output.writeString(1, getName());
+        if (((bitField0_ & 0x00000001) == 0x00000001)) {
+          output.writeBytes(1, getNameBytes());
         }
-        if (hasStartCode()) {
-          output.writeInt64(2, getStartCode());
+        if (((bitField0_ & 0x00000002) == 0x00000002)) {
+          output.writeInt64(2, startCode_);
         }
-        if (hasRequests()) {
-          output.writeInt32(3, getRequests());
+        if (((bitField0_ & 0x00000004) == 0x00000004)) {
+          output.writeInt32(3, requests_);
         }
-        if (hasHeapSizeMB()) {
-          output.writeInt32(4, getHeapSizeMB());
+        if (((bitField0_ & 0x00000008) == 0x00000008)) {
+          output.writeInt32(4, heapSizeMB_);
         }
-        if (hasMaxHeapSizeMB()) {
-          output.writeInt32(5, getMaxHeapSizeMB());
+        if (((bitField0_ & 0x00000010) == 0x00000010)) {
+          output.writeInt32(5, maxHeapSizeMB_);
         }
-        for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
-          output.writeMessage(6, element);
+        for (int i = 0; i < regions_.size(); i++) {
+          output.writeMessage(6, regions_.get(i));
         }
         getUnknownFields().writeTo(output);
       }
@@ -632,35 +912,42 @@ public final class StorageClusterStatusM
         if (size != -1) return size;
       
         size = 0;
-        if (hasName()) {
+        if (((bitField0_ & 0x00000001) == 0x00000001)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeStringSize(1, getName());
+            .computeBytesSize(1, getNameBytes());
         }
-        if (hasStartCode()) {
+        if (((bitField0_ & 0x00000002) == 0x00000002)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt64Size(2, getStartCode());
+            .computeInt64Size(2, startCode_);
         }
-        if (hasRequests()) {
+        if (((bitField0_ & 0x00000004) == 0x00000004)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(3, getRequests());
+            .computeInt32Size(3, requests_);
         }
-        if (hasHeapSizeMB()) {
+        if (((bitField0_ & 0x00000008) == 0x00000008)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(4, getHeapSizeMB());
+            .computeInt32Size(4, heapSizeMB_);
         }
-        if (hasMaxHeapSizeMB()) {
+        if (((bitField0_ & 0x00000010) == 0x00000010)) {
           size += com.google.protobuf.CodedOutputStream
-            .computeInt32Size(5, getMaxHeapSizeMB());
+            .computeInt32Size(5, maxHeapSizeMB_);
         }
-        for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region element : getRegionsList()) {
+        for (int i = 0; i < regions_.size(); i++) {
           size += com.google.protobuf.CodedOutputStream
-            .computeMessageSize(6, element);
+            .computeMessageSize(6, regions_.get(i));
         }
         size += getUnknownFields().getSerializedSize();
         memoizedSerializedSize = size;
         return size;
       }
       
+      private static final long serialVersionUID = 0L;
+      @java.lang.Override
+      protected java.lang.Object writeReplace()
+          throws java.io.ObjectStreamException {
+        return super.writeReplace();
+      }
+      
       public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node parseFrom(
           com.google.protobuf.ByteString data)
           throws com.google.protobuf.InvalidProtocolBufferException {
@@ -735,34 +1022,66 @@ public final class StorageClusterStatusM
       }
       public Builder toBuilder() { return newBuilder(this); }
       
+      @java.lang.Override
+      protected Builder newBuilderForType(
+          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        Builder builder = new Builder(parent);
+        return builder;
+      }
       public static final class Builder extends
-          com.google.protobuf.GeneratedMessage.Builder<Builder> {
-        private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node result;
+          com.google.protobuf.GeneratedMessage.Builder<Builder>
+         implements org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder {
+        public static final com.google.protobuf.Descriptors.Descriptor
+            getDescriptor() {
+          return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_descriptor;
+        }
         
-        // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.newBuilder()
-        private Builder() {}
+        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+            internalGetFieldAccessorTable() {
+          return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_Node_fieldAccessorTable;
+        }
         
-        private static Builder create() {
-          Builder builder = new Builder();
-          builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node();
-          return builder;
+        // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.newBuilder()
+        private Builder() {
+          maybeForceBuilderInitialization();
         }
         
-        protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node internalGetResult() {
-          return result;
+        private Builder(BuilderParent parent) {
+          super(parent);
+          maybeForceBuilderInitialization();
+        }
+        private void maybeForceBuilderInitialization() {
+          if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+            getRegionsFieldBuilder();
+          }
+        }
+        private static Builder create() {
+          return new Builder();
         }
         
         public Builder clear() {
-          if (result == null) {
-            throw new IllegalStateException(
-              "Cannot call clear() after build().");
+          super.clear();
+          name_ = "";
+          bitField0_ = (bitField0_ & ~0x00000001);
+          startCode_ = 0L;
+          bitField0_ = (bitField0_ & ~0x00000002);
+          requests_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000004);
+          heapSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000008);
+          maxHeapSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000010);
+          if (regionsBuilder_ == null) {
+            regions_ = java.util.Collections.emptyList();
+            bitField0_ = (bitField0_ & ~0x00000020);
+          } else {
+            regionsBuilder_.clear();
           }
-          result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node();
           return this;
         }
         
         public Builder clone() {
-          return create().mergeFrom(result);
+          return create().mergeFrom(buildPartial());
         }
         
         public com.google.protobuf.Descriptors.Descriptor
@@ -774,37 +1093,60 @@ public final class StorageClusterStatusM
           return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.getDefaultInstance();
         }
         
-        public boolean isInitialized() {
-          return result.isInitialized();
-        }
         public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node build() {
-          if (result != null && !isInitialized()) {
+          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node result = buildPartial();
+          if (!result.isInitialized()) {
             throw newUninitializedMessageException(result);
           }
-          return buildPartial();
+          return result;
         }
         
         private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node buildParsed()
             throws com.google.protobuf.InvalidProtocolBufferException {
-          if (!isInitialized()) {
+          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node result = buildPartial();
+          if (!result.isInitialized()) {
             throw newUninitializedMessageException(
               result).asInvalidProtocolBufferException();
           }
-          return buildPartial();
+          return result;
         }
         
         public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node buildPartial() {
-          if (result == null) {
-            throw new IllegalStateException(
-              "build() has already been called on this Builder.");
-          }
-          if (result.regions_ != java.util.Collections.EMPTY_LIST) {
-            result.regions_ =
-              java.util.Collections.unmodifiableList(result.regions_);
-          }
-          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node returnMe = result;
-          result = null;
-          return returnMe;
+          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node(this);
+          int from_bitField0_ = bitField0_;
+          int to_bitField0_ = 0;
+          if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+            to_bitField0_ |= 0x00000001;
+          }
+          result.name_ = name_;
+          if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+            to_bitField0_ |= 0x00000002;
+          }
+          result.startCode_ = startCode_;
+          if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+            to_bitField0_ |= 0x00000004;
+          }
+          result.requests_ = requests_;
+          if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
+            to_bitField0_ |= 0x00000008;
+          }
+          result.heapSizeMB_ = heapSizeMB_;
+          if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
+            to_bitField0_ |= 0x00000010;
+          }
+          result.maxHeapSizeMB_ = maxHeapSizeMB_;
+          if (regionsBuilder_ == null) {
+            if (((bitField0_ & 0x00000020) == 0x00000020)) {
+              regions_ = java.util.Collections.unmodifiableList(regions_);
+              bitField0_ = (bitField0_ & ~0x00000020);
+            }
+            result.regions_ = regions_;
+          } else {
+            result.regions_ = regionsBuilder_.build();
+          }
+          result.bitField0_ = to_bitField0_;
+          onBuilt();
+          return result;
         }
         
         public Builder mergeFrom(com.google.protobuf.Message other) {
@@ -833,16 +1175,50 @@ public final class StorageClusterStatusM
           if (other.hasMaxHeapSizeMB()) {
             setMaxHeapSizeMB(other.getMaxHeapSizeMB());
           }
-          if (!other.regions_.isEmpty()) {
-            if (result.regions_.isEmpty()) {
-              result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
+          if (regionsBuilder_ == null) {
+            if (!other.regions_.isEmpty()) {
+              if (regions_.isEmpty()) {
+                regions_ = other.regions_;
+                bitField0_ = (bitField0_ & ~0x00000020);
+              } else {
+                ensureRegionsIsMutable();
+                regions_.addAll(other.regions_);
+              }
+              onChanged();
+            }
+          } else {
+            if (!other.regions_.isEmpty()) {
+              if (regionsBuilder_.isEmpty()) {
+                regionsBuilder_.dispose();
+                regionsBuilder_ = null;
+                regions_ = other.regions_;
+                bitField0_ = (bitField0_ & ~0x00000020);
+                regionsBuilder_ = 
+                  com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+                     getRegionsFieldBuilder() : null;
+              } else {
+                regionsBuilder_.addAllMessages(other.regions_);
+              }
             }
-            result.regions_.addAll(other.regions_);
           }
           this.mergeUnknownFields(other.getUnknownFields());
           return this;
         }
         
+        public final boolean isInitialized() {
+          if (!hasName()) {
+            
+            return false;
+          }
+          for (int i = 0; i < getRegionsCount(); i++) {
+            if (!getRegions(i).isInitialized()) {
+              
+              return false;
+            }
+          }
+          return true;
+        }
+        
         public Builder mergeFrom(
             com.google.protobuf.CodedInputStream input,
             com.google.protobuf.ExtensionRegistryLite extensionRegistry)
@@ -855,33 +1231,40 @@ public final class StorageClusterStatusM
             switch (tag) {
               case 0:
                 this.setUnknownFields(unknownFields.build());
+                onChanged();
                 return this;
               default: {
                 if (!parseUnknownField(input, unknownFields,
                                        extensionRegistry, tag)) {
                   this.setUnknownFields(unknownFields.build());
+                  onChanged();
                   return this;
                 }
                 break;
               }
               case 10: {
-                setName(input.readString());
+                bitField0_ |= 0x00000001;
+                name_ = input.readBytes();
                 break;
               }
               case 16: {
-                setStartCode(input.readInt64());
+                bitField0_ |= 0x00000002;
+                startCode_ = input.readInt64();
                 break;
               }
               case 24: {
-                setRequests(input.readInt32());
+                bitField0_ |= 0x00000004;
+                requests_ = input.readInt32();
                 break;
               }
               case 32: {
-                setHeapSizeMB(input.readInt32());
+                bitField0_ |= 0x00000008;
+                heapSizeMB_ = input.readInt32();
                 break;
               }
               case 40: {
-                setMaxHeapSizeMB(input.readInt32());
+                bitField0_ |= 0x00000010;
+                maxHeapSizeMB_ = input.readInt32();
                 break;
               }
               case 50: {
@@ -894,234 +1277,430 @@ public final class StorageClusterStatusM
           }
         }
         
+        private int bitField0_;
         
         // required string name = 1;
+        private java.lang.Object name_ = "";
         public boolean hasName() {
-          return result.hasName();
+          return ((bitField0_ & 0x00000001) == 0x00000001);
         }
-        public java.lang.String getName() {
-          return result.getName();
+        public String getName() {
+          java.lang.Object ref = name_;
+          if (!(ref instanceof String)) {
+            String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
+            name_ = s;
+            return s;
+          } else {
+            return (String) ref;
+          }
         }
-        public Builder setName(java.lang.String value) {
+        public Builder setName(String value) {
           if (value == null) {
     throw new NullPointerException();
   }
-  result.hasName = true;
-          result.name_ = value;
+  bitField0_ |= 0x00000001;
+          name_ = value;
+          onChanged();
           return this;
         }
         public Builder clearName() {
-          result.hasName = false;
-          result.name_ = getDefaultInstance().getName();
+          bitField0_ = (bitField0_ & ~0x00000001);
+          name_ = getDefaultInstance().getName();
+          onChanged();
           return this;
         }
+        void setName(com.google.protobuf.ByteString value) {
+          bitField0_ |= 0x00000001;
+          name_ = value;
+          onChanged();
+        }
         
         // optional int64 startCode = 2;
+        private long startCode_ ;
         public boolean hasStartCode() {
-          return result.hasStartCode();
+          return ((bitField0_ & 0x00000002) == 0x00000002);
         }
         public long getStartCode() {
-          return result.getStartCode();
+          return startCode_;
         }
         public Builder setStartCode(long value) {
-          result.hasStartCode = true;
-          result.startCode_ = value;
+          bitField0_ |= 0x00000002;
+          startCode_ = value;
+          onChanged();
           return this;
         }
         public Builder clearStartCode() {
-          result.hasStartCode = false;
-          result.startCode_ = 0L;
+          bitField0_ = (bitField0_ & ~0x00000002);
+          startCode_ = 0L;
+          onChanged();
           return this;
         }
         
         // optional int32 requests = 3;
+        private int requests_ ;
         public boolean hasRequests() {
-          return result.hasRequests();
+          return ((bitField0_ & 0x00000004) == 0x00000004);
         }
         public int getRequests() {
-          return result.getRequests();
+          return requests_;
         }
         public Builder setRequests(int value) {
-          result.hasRequests = true;
-          result.requests_ = value;
+          bitField0_ |= 0x00000004;
+          requests_ = value;
+          onChanged();
           return this;
         }
         public Builder clearRequests() {
-          result.hasRequests = false;
-          result.requests_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000004);
+          requests_ = 0;
+          onChanged();
           return this;
         }
         
         // optional int32 heapSizeMB = 4;
+        private int heapSizeMB_ ;
         public boolean hasHeapSizeMB() {
-          return result.hasHeapSizeMB();
+          return ((bitField0_ & 0x00000008) == 0x00000008);
         }
         public int getHeapSizeMB() {
-          return result.getHeapSizeMB();
+          return heapSizeMB_;
         }
         public Builder setHeapSizeMB(int value) {
-          result.hasHeapSizeMB = true;
-          result.heapSizeMB_ = value;
+          bitField0_ |= 0x00000008;
+          heapSizeMB_ = value;
+          onChanged();
           return this;
         }
         public Builder clearHeapSizeMB() {
-          result.hasHeapSizeMB = false;
-          result.heapSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000008);
+          heapSizeMB_ = 0;
+          onChanged();
           return this;
         }
         
         // optional int32 maxHeapSizeMB = 5;
+        private int maxHeapSizeMB_ ;
         public boolean hasMaxHeapSizeMB() {
-          return result.hasMaxHeapSizeMB();
+          return ((bitField0_ & 0x00000010) == 0x00000010);
         }
         public int getMaxHeapSizeMB() {
-          return result.getMaxHeapSizeMB();
+          return maxHeapSizeMB_;
         }
         public Builder setMaxHeapSizeMB(int value) {
-          result.hasMaxHeapSizeMB = true;
-          result.maxHeapSizeMB_ = value;
+          bitField0_ |= 0x00000010;
+          maxHeapSizeMB_ = value;
+          onChanged();
           return this;
         }
         public Builder clearMaxHeapSizeMB() {
-          result.hasMaxHeapSizeMB = false;
-          result.maxHeapSizeMB_ = 0;
+          bitField0_ = (bitField0_ & ~0x00000010);
+          maxHeapSizeMB_ = 0;
+          onChanged();
           return this;
         }
         
         // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Region regions = 6;
+        private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> regions_ =
+          java.util.Collections.emptyList();
+        private void ensureRegionsIsMutable() {
+          if (!((bitField0_ & 0x00000020) == 0x00000020)) {
+            regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>(regions_);
+            bitField0_ |= 0x00000020;
+           }
+        }
+        
+        private com.google.protobuf.RepeatedFieldBuilder<
+            org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder> regionsBuilder_;
+        
         public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> getRegionsList() {
-          return java.util.Collections.unmodifiableList(result.regions_);
+          if (regionsBuilder_ == null) {
+            return java.util.Collections.unmodifiableList(regions_);
+          } else {
+            return regionsBuilder_.getMessageList();
+          }
         }
         public int getRegionsCount() {
-          return result.getRegionsCount();
+          if (regionsBuilder_ == null) {
+            return regions_.size();
+          } else {
+            return regionsBuilder_.getCount();
+          }
         }
         public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region getRegions(int index) {
-          return result.getRegions(index);
+          if (regionsBuilder_ == null) {
+            return regions_.get(index);
+          } else {
+            return regionsBuilder_.getMessage(index);
+          }
         }
-        public Builder setRegions(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) {
-          if (value == null) {
-            throw new NullPointerException();
+        public Builder setRegions(
+            int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) {
+          if (regionsBuilder_ == null) {
+            if (value == null) {
+              throw new NullPointerException();
+            }
+            ensureRegionsIsMutable();
+            regions_.set(index, value);
+            onChanged();
+          } else {
+            regionsBuilder_.setMessage(index, value);
           }
-          result.regions_.set(index, value);
           return this;
         }
-        public Builder setRegions(int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
-          result.regions_.set(index, builderForValue.build());
+        public Builder setRegions(
+            int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
+          if (regionsBuilder_ == null) {
+            ensureRegionsIsMutable();
+            regions_.set(index, builderForValue.build());
+            onChanged();
+          } else {
+            regionsBuilder_.setMessage(index, builderForValue.build());
+          }
           return this;
         }
         public Builder addRegions(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) {
-          if (value == null) {
-            throw new NullPointerException();
+          if (regionsBuilder_ == null) {
+            if (value == null) {
+              throw new NullPointerException();
+            }
+            ensureRegionsIsMutable();
+            regions_.add(value);
+            onChanged();
+          } else {
+            regionsBuilder_.addMessage(value);
           }
-          if (result.regions_.isEmpty()) {
-            result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
+          return this;
+        }
+        public Builder addRegions(
+            int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region value) {
+          if (regionsBuilder_ == null) {
+            if (value == null) {
+              throw new NullPointerException();
+            }
+            ensureRegionsIsMutable();
+            regions_.add(index, value);
+            onChanged();
+          } else {
+            regionsBuilder_.addMessage(index, value);
           }
-          result.regions_.add(value);
           return this;
         }
-        public Builder addRegions(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
-          if (result.regions_.isEmpty()) {
-            result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
+        public Builder addRegions(
+            org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
+          if (regionsBuilder_ == null) {
+            ensureRegionsIsMutable();
+            regions_.add(builderForValue.build());
+            onChanged();
+          } else {
+            regionsBuilder_.addMessage(builderForValue.build());
+          }
+          return this;
+        }
+        public Builder addRegions(
+            int index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder builderForValue) {
+          if (regionsBuilder_ == null) {
+            ensureRegionsIsMutable();
+            regions_.add(index, builderForValue.build());
+            onChanged();
+          } else {
+            regionsBuilder_.addMessage(index, builderForValue.build());
           }
-          result.regions_.add(builderForValue.build());
           return this;
         }
         public Builder addAllRegions(
             java.lang.Iterable<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region> values) {
-          if (result.regions_.isEmpty()) {
-            result.regions_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region>();
+          if (regionsBuilder_ == null) {
+            ensureRegionsIsMutable();
+            super.addAll(values, regions_);
+            onChanged();
+          } else {
+            regionsBuilder_.addAllMessages(values);
           }
-          super.addAll(values, result.regions_);
           return this;
         }
         public Builder clearRegions() {
-          result.regions_ = java.util.Collections.emptyList();
+          if (regionsBuilder_ == null) {
+            regions_ = java.util.Collections.emptyList();
+            bitField0_ = (bitField0_ & ~0x00000020);
+            onChanged();
+          } else {
+            regionsBuilder_.clear();
+          }
           return this;
         }
+        public Builder removeRegions(int index) {
+          if (regionsBuilder_ == null) {
+            ensureRegionsIsMutable();
+            regions_.remove(index);
+            onChanged();
+          } else {
+            regionsBuilder_.remove(index);
+          }
+          return this;
+        }
+        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder getRegionsBuilder(
+            int index) {
+          return getRegionsFieldBuilder().getBuilder(index);
+        }
+        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder getRegionsOrBuilder(
+            int index) {
+          if (regionsBuilder_ == null) {
+            return regions_.get(index);  } else {
+            return regionsBuilder_.getMessageOrBuilder(index);
+          }
+        }
+        public java.util.List<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder> 
+             getRegionsOrBuilderList() {
+          if (regionsBuilder_ != null) {
+            return regionsBuilder_.getMessageOrBuilderList();
+          } else {
+            return java.util.Collections.unmodifiableList(regions_);
+          }
+        }
+        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder addRegionsBuilder() {
+          return getRegionsFieldBuilder().addBuilder(
+              org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance());
+        }
+        public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder addRegionsBuilder(
+            int index) {
+          return getRegionsFieldBuilder().addBuilder(
+              index, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.getDefaultInstance());
+        }
+        public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder> 
+             getRegionsBuilderList() {
+          return getRegionsFieldBuilder().getBuilderList();
+        }
+        private com.google.protobuf.RepeatedFieldBuilder<
+            org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder> 
+            getRegionsFieldBuilder() {
+          if (regionsBuilder_ == null) {
+            regionsBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+                org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Region.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.RegionOrBuilder>(
+                    regions_,
+                    ((bitField0_ & 0x00000020) == 0x00000020),
+                    getParentForChildren(),
+                    isClean());
+            regions_ = null;
+          }
+          return regionsBuilder_;
+        }
         
         // @@protoc_insertion_point(builder_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node)
       }
       
       static {
         defaultInstance = new Node(true);
-        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internalForceInit();
         defaultInstance.initFields();
       }
       
       // @@protoc_insertion_point(class_scope:org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node)
     }
     
+    private int bitField0_;
     // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1;
     public static final int LIVENODES_FIELD_NUMBER = 1;
-    private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> liveNodes_ =
-      java.util.Collections.emptyList();
+    private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> liveNodes_;
     public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> getLiveNodesList() {
       return liveNodes_;
     }
-    public int getLiveNodesCount() { return liveNodes_.size(); }
+    public java.util.List<? extends org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder> 
+        getLiveNodesOrBuilderList() {
+      return liveNodes_;
+    }
+    public int getLiveNodesCount() {
+      return liveNodes_.size();
+    }
     public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index) {
       return liveNodes_.get(index);
     }
+    public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder getLiveNodesOrBuilder(
+        int index) {
+      return liveNodes_.get(index);
+    }
     
     // repeated string deadNodes = 2;
     public static final int DEADNODES_FIELD_NUMBER = 2;
-    private java.util.List<java.lang.String> deadNodes_ =
-      java.util.Collections.emptyList();
-    public java.util.List<java.lang.String> getDeadNodesList() {
+    private com.google.protobuf.LazyStringList deadNodes_;
+    public java.util.List<String>
+        getDeadNodesList() {
       return deadNodes_;
     }
-    public int getDeadNodesCount() { return deadNodes_.size(); }
-    public java.lang.String getDeadNodes(int index) {
+    public int getDeadNodesCount() {
+      return deadNodes_.size();
+    }
+    public String getDeadNodes(int index) {
       return deadNodes_.get(index);
     }
     
     // optional int32 regions = 3;
     public static final int REGIONS_FIELD_NUMBER = 3;
-    private boolean hasRegions;
-    private int regions_ = 0;
-    public boolean hasRegions() { return hasRegions; }
-    public int getRegions() { return regions_; }
+    private int regions_;
+    public boolean hasRegions() {
+      return ((bitField0_ & 0x00000001) == 0x00000001);
+    }
+    public int getRegions() {
+      return regions_;
+    }
     
     // optional int32 requests = 4;
     public static final int REQUESTS_FIELD_NUMBER = 4;
-    private boolean hasRequests;
-    private int requests_ = 0;
-    public boolean hasRequests() { return hasRequests; }
-    public int getRequests() { return requests_; }
+    private int requests_;
+    public boolean hasRequests() {
+      return ((bitField0_ & 0x00000002) == 0x00000002);
+    }
+    public int getRequests() {
+      return requests_;
+    }
     
     // optional double averageLoad = 5;
     public static final int AVERAGELOAD_FIELD_NUMBER = 5;
-    private boolean hasAverageLoad;
-    private double averageLoad_ = 0D;
-    public boolean hasAverageLoad() { return hasAverageLoad; }
-    public double getAverageLoad() { return averageLoad_; }
+    private double averageLoad_;
+    public boolean hasAverageLoad() {
+      return ((bitField0_ & 0x00000004) == 0x00000004);
+    }
+    public double getAverageLoad() {
+      return averageLoad_;
+    }
     
     private void initFields() {
+      liveNodes_ = java.util.Collections.emptyList();
+      deadNodes_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+      regions_ = 0;
+      requests_ = 0;
+      averageLoad_ = 0D;
     }
+    private byte memoizedIsInitialized = -1;
     public final boolean isInitialized() {
-      for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
-        if (!element.isInitialized()) return false;
+      byte isInitialized = memoizedIsInitialized;
+      if (isInitialized != -1) return isInitialized == 1;
+      
+      for (int i = 0; i < getLiveNodesCount(); i++) {
+        if (!getLiveNodes(i).isInitialized()) {
+          memoizedIsInitialized = 0;
+          return false;
+        }
       }
+      memoizedIsInitialized = 1;
       return true;
     }
     
     public void writeTo(com.google.protobuf.CodedOutputStream output)
                         throws java.io.IOException {
       getSerializedSize();
-      for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
-        output.writeMessage(1, element);
+      for (int i = 0; i < liveNodes_.size(); i++) {
+        output.writeMessage(1, liveNodes_.get(i));
       }
-      for (java.lang.String element : getDeadNodesList()) {
-        output.writeString(2, element);
+      for (int i = 0; i < deadNodes_.size(); i++) {
+        output.writeBytes(2, deadNodes_.getByteString(i));
       }
-      if (hasRegions()) {
-        output.writeInt32(3, getRegions());
+      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+        output.writeInt32(3, regions_);
       }
-      if (hasRequests()) {
-        output.writeInt32(4, getRequests());
+      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+        output.writeInt32(4, requests_);
       }
-      if (hasAverageLoad()) {
-        output.writeDouble(5, getAverageLoad());
+      if (((bitField0_ & 0x00000004) == 0x00000004)) {
+        output.writeDouble(5, averageLoad_);
       }
       getUnknownFields().writeTo(output);
     }
@@ -1132,36 +1711,43 @@ public final class StorageClusterStatusM
       if (size != -1) return size;
     
       size = 0;
-      for (org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node element : getLiveNodesList()) {
+      for (int i = 0; i < liveNodes_.size(); i++) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, element);
+          .computeMessageSize(1, liveNodes_.get(i));
       }
       {
         int dataSize = 0;
-        for (java.lang.String element : getDeadNodesList()) {
+        for (int i = 0; i < deadNodes_.size(); i++) {
           dataSize += com.google.protobuf.CodedOutputStream
-            .computeStringSizeNoTag(element);
+            .computeBytesSizeNoTag(deadNodes_.getByteString(i));
         }
         size += dataSize;
         size += 1 * getDeadNodesList().size();
       }
-      if (hasRegions()) {
+      if (((bitField0_ & 0x00000001) == 0x00000001)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeInt32Size(3, getRegions());
+          .computeInt32Size(3, regions_);
       }
-      if (hasRequests()) {
+      if (((bitField0_ & 0x00000002) == 0x00000002)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeInt32Size(4, getRequests());
+          .computeInt32Size(4, requests_);
       }
-      if (hasAverageLoad()) {
+      if (((bitField0_ & 0x00000004) == 0x00000004)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeDoubleSize(5, getAverageLoad());
+          .computeDoubleSize(5, averageLoad_);
       }
       size += getUnknownFields().getSerializedSize();
       memoizedSerializedSize = size;
       return size;
     }
     
+    private static final long serialVersionUID = 0L;
+    @java.lang.Override
+    protected java.lang.Object writeReplace()
+        throws java.io.ObjectStreamException {
+      return super.writeReplace();
+    }
+    
     public static org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -1236,34 +1822,64 @@ public final class StorageClusterStatusM
     }
     public Builder toBuilder() { return newBuilder(this); }
     
+    @java.lang.Override
+    protected Builder newBuilderForType(
+        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+      Builder builder = new Builder(parent);
+      return builder;
+    }
     public static final class Builder extends
-        com.google.protobuf.GeneratedMessage.Builder<Builder> {
-      private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus result;
+        com.google.protobuf.GeneratedMessage.Builder<Builder>
+       implements org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatusOrBuilder {
+      public static final com.google.protobuf.Descriptors.Descriptor
+          getDescriptor() {
+        return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_descriptor;
+      }
       
-      // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.newBuilder()
-      private Builder() {}
+      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+          internalGetFieldAccessorTable() {
+        return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.internal_static_org_apache_hadoop_hbase_rest_protobuf_generated_StorageClusterStatus_fieldAccessorTable;
+      }
       
-      private static Builder create() {
-        Builder builder = new Builder();
-        builder.result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus();
-        return builder;
+      // Construct using org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.newBuilder()
+      private Builder() {
+        maybeForceBuilderInitialization();
       }
       
-      protected org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus internalGetResult() {
-        return result;
+      private Builder(BuilderParent parent) {
+        super(parent);
+        maybeForceBuilderInitialization();
+      }
+      private void maybeForceBuilderInitialization() {
+        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+          getLiveNodesFieldBuilder();
+        }
+      }
+      private static Builder create() {
+        return new Builder();
       }
       
       public Builder clear() {
-        if (result == null) {
-          throw new IllegalStateException(
-            "Cannot call clear() after build().");
+        super.clear();
+        if (liveNodesBuilder_ == null) {
+          liveNodes_ = java.util.Collections.emptyList();
+          bitField0_ = (bitField0_ & ~0x00000001);
+        } else {
+          liveNodesBuilder_.clear();
         }
-        result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus();
+        deadNodes_ = com.google.protobuf.LazyStringArrayList.EMPTY;
+        bitField0_ = (bitField0_ & ~0x00000002);
+        regions_ = 0;
+        bitField0_ = (bitField0_ & ~0x00000004);
+        requests_ = 0;
+        bitField0_ = (bitField0_ & ~0x00000008);
+        averageLoad_ = 0D;
+        bitField0_ = (bitField0_ & ~0x00000010);
         return this;
       }
       
       public Builder clone() {
-        return create().mergeFrom(result);
+        return create().mergeFrom(buildPartial());
       }
       
       public com.google.protobuf.Descriptors.Descriptor
@@ -1275,41 +1891,58 @@ public final class StorageClusterStatusM
         return org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDefaultInstance();
       }
       
-      public boolean isInitialized() {
-        return result.isInitialized();
-      }
       public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus build() {
-        if (result != null && !isInitialized()) {
+        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus result = buildPartial();
+        if (!result.isInitialized()) {
           throw newUninitializedMessageException(result);
         }
-        return buildPartial();
+        return result;
       }
       
       private org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus buildParsed()
           throws com.google.protobuf.InvalidProtocolBufferException {
-        if (!isInitialized()) {
+        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus result = buildPartial();
+        if (!result.isInitialized()) {
           throw newUninitializedMessageException(
             result).asInvalidProtocolBufferException();
         }
-        return buildPartial();
+        return result;
       }
       
       public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus buildPartial() {
-        if (result == null) {
-          throw new IllegalStateException(
-            "build() has already been called on this Builder.");
-        }
-        if (result.liveNodes_ != java.util.Collections.EMPTY_LIST) {
-          result.liveNodes_ =
-            java.util.Collections.unmodifiableList(result.liveNodes_);
-        }
-        if (result.deadNodes_ != java.util.Collections.EMPTY_LIST) {
-          result.deadNodes_ =
-            java.util.Collections.unmodifiableList(result.deadNodes_);
-        }
-        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus returnMe = result;
-        result = null;
-        return returnMe;
+        org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus result = new org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus(this);
+        int from_bitField0_ = bitField0_;
+        int to_bitField0_ = 0;
+        if (liveNodesBuilder_ == null) {
+          if (((bitField0_ & 0x00000001) == 0x00000001)) {
+            liveNodes_ = java.util.Collections.unmodifiableList(liveNodes_);
+            bitField0_ = (bitField0_ & ~0x00000001);
+          }
+          result.liveNodes_ = liveNodes_;
+        } else {
+          result.liveNodes_ = liveNodesBuilder_.build();
+        }
+        if (((bitField0_ & 0x00000002) == 0x00000002)) {
+          deadNodes_ = new com.google.protobuf.UnmodifiableLazyStringList(
+              deadNodes_);
+          bitField0_ = (bitField0_ & ~0x00000002);
+        }
+        result.deadNodes_ = deadNodes_;
+        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
+          to_bitField0_ |= 0x00000001;
+        }
+        result.regions_ = regions_;
+        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
+          to_bitField0_ |= 0x00000002;
+        }
+        result.requests_ = requests_;
+        if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
+          to_bitField0_ |= 0x00000004;
+        }
+        result.averageLoad_ = averageLoad_;
+        result.bitField0_ = to_bitField0_;
+        onBuilt();
+        return result;
       }
       
       public Builder mergeFrom(com.google.protobuf.Message other) {
@@ -1323,17 +1956,41 @@ public final class StorageClusterStatusM
       
       public Builder mergeFrom(org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus other) {
         if (other == org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.getDefaultInstance()) return this;
-        if (!other.liveNodes_.isEmpty()) {
-          if (result.liveNodes_.isEmpty()) {
-            result.liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>();
+        if (liveNodesBuilder_ == null) {
+          if (!other.liveNodes_.isEmpty()) {
+            if (liveNodes_.isEmpty()) {
+              liveNodes_ = other.liveNodes_;
+              bitField0_ = (bitField0_ & ~0x00000001);
+            } else {
+              ensureLiveNodesIsMutable();
+              liveNodes_.addAll(other.liveNodes_);
+            }
+            onChanged();
+          }
+        } else {
+          if (!other.liveNodes_.isEmpty()) {
+            if (liveNodesBuilder_.isEmpty()) {
+              liveNodesBuilder_.dispose();
+              liveNodesBuilder_ = null;
+              liveNodes_ = other.liveNodes_;
+              bitField0_ = (bitField0_ & ~0x00000001);
+              liveNodesBuilder_ = 
+                com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+                   getLiveNodesFieldBuilder() : null;
+            } else {
+              liveNodesBuilder_.addAllMessages(other.liveNodes_);
+            }
           }
-          result.liveNodes_.addAll(other.liveNodes_);
         }
         if (!other.deadNodes_.isEmpty()) {
-          if (result.deadNodes_.isEmpty()) {
-            result.deadNodes_ = new java.util.ArrayList<java.lang.String>();
+          if (deadNodes_.isEmpty()) {
+            deadNodes_ = other.deadNodes_;
+            bitField0_ = (bitField0_ & ~0x00000002);
+          } else {
+            ensureDeadNodesIsMutable();
+            deadNodes_.addAll(other.deadNodes_);
           }
-          result.deadNodes_.addAll(other.deadNodes_);
+          onChanged();
         }
         if (other.hasRegions()) {
           setRegions(other.getRegions());
@@ -1348,6 +2005,16 @@ public final class StorageClusterStatusM
         return this;
       }
       
+      public final boolean isInitialized() {
+        for (int i = 0; i < getLiveNodesCount(); i++) {
+          if (!getLiveNodes(i).isInitialized()) {
+            
+            return false;
+          }
+        }
+        return true;
+      }
+      
       public Builder mergeFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
@@ -1360,11 +2027,13 @@ public final class StorageClusterStatusM
           switch (tag) {
             case 0:
               this.setUnknownFields(unknownFields.build());
+              onChanged();
               return this;
             default: {
               if (!parseUnknownField(input, unknownFields,
                                      extensionRegistry, tag)) {
                 this.setUnknownFields(unknownFields.build());
+                onChanged();
                 return this;
               }
               break;
@@ -1376,168 +2045,333 @@ public final class StorageClusterStatusM
               break;
             }
             case 18: {
-              addDeadNodes(input.readString());
+              ensureDeadNodesIsMutable();
+              deadNodes_.add(input.readBytes());
               break;
             }
             case 24: {
-              setRegions(input.readInt32());
+              bitField0_ |= 0x00000004;
+              regions_ = input.readInt32();
               break;
             }
             case 32: {
-              setRequests(input.readInt32());
+              bitField0_ |= 0x00000008;
+              requests_ = input.readInt32();
               break;
             }
             case 41: {
-              setAverageLoad(input.readDouble());
+              bitField0_ |= 0x00000010;
+              averageLoad_ = input.readDouble();
               break;
             }
           }
         }
       }
       
+      private int bitField0_;
       
       // repeated .org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatus.Node liveNodes = 1;
+      private java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> liveNodes_ =
+        java.util.Collections.emptyList();
+      private void ensureLiveNodesIsMutable() {
+        if (!((bitField0_ & 0x00000001) == 0x00000001)) {
+          liveNodes_ = new java.util.ArrayList<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node>(liveNodes_);
+          bitField0_ |= 0x00000001;
+         }
+      }
+      
+      private com.google.protobuf.RepeatedFieldBuilder<
+          org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node.Builder, org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.NodeOrBuilder> liveNodesBuilder_;
+      
       public java.util.List<org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node> getLiveNodesList() {
-        return java.util.Collections.unmodifiableList(result.liveNodes_);
+        if (liveNodesBuilder_ == null) {
+          return java.util.Collections.unmodifiableList(liveNodes_);
+        } else {
+          return liveNodesBuilder_.getMessageList();
+        }
       }
       public int getLiveNodesCount() {
-        return result.getLiveNodesCount();
+        if (liveNodesBuilder_ == null) {
+          return liveNodes_.size();
+        } else {
+          return liveNodesBuilder_.getCount();
+        }
       }
       public org.apache.hadoop.hbase.rest.protobuf.generated.StorageClusterStatusMessage.StorageClusterStatus.Node getLiveNodes(int index) {
-        return result.getLiveNodes(index);
+        if (liveNodesBuilder_ == null) {

[... 348 lines stripped ...]