You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2013/05/03 05:58:35 UTC

svn commit: r1478639 [6/10] - in /hbase/branches/0.95: hbase-client/src/main/java/org/apache/hadoop/hbase/ hbase-client/src/main/java/org/apache/hadoop/hbase/catalog/ hbase-client/src/main/java/org/apache/hadoop/hbase/client/ hbase-client/src/main/java...

Modified: hbase/branches/0.95/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java (original)
+++ hbase/branches/0.95/hbase-client/src/test/java/org/apache/hadoop/hbase/client/TestSnapshotFromAdmin.java Fri May  3 03:58:33 2013
@@ -78,7 +78,7 @@ public class TestSnapshotFromAdmin {
     // mock the master admin to our mock
     MasterAdminKeepAliveConnection mockMaster = Mockito.mock(MasterAdminKeepAliveConnection.class);
     Mockito.when(mockConnection.getConfiguration()).thenReturn(conf);
-    Mockito.when(mockConnection.getKeepAliveMasterAdmin()).thenReturn(mockMaster);
+    Mockito.when(mockConnection.getKeepAliveMasterAdminService()).thenReturn(mockMaster);
     // set the max wait time for the snapshot to complete
     TakeSnapshotResponse response = TakeSnapshotResponse.newBuilder()
         .setExpectedTimeout(maxWaitTime)
@@ -135,7 +135,7 @@ public class TestSnapshotFromAdmin {
 
     // mock the master connection
     MasterAdminKeepAliveConnection master = Mockito.mock(MasterAdminKeepAliveConnection.class);
-    Mockito.when(mockConnection.getKeepAliveMasterAdmin()).thenReturn(master);
+    Mockito.when(mockConnection.getKeepAliveMasterAdminService()).thenReturn(master);
     TakeSnapshotResponse response = TakeSnapshotResponse.newBuilder().setExpectedTimeout(0).build();
     Mockito.when(
       master.snapshot((RpcController) Mockito.isNull(), Mockito.any(TakeSnapshotRequest.class)))

Modified: hbase/branches/0.95/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java (original)
+++ hbase/branches/0.95/hbase-common/src/main/java/org/apache/hadoop/hbase/HConstants.java Fri May  3 03:58:33 2013
@@ -573,17 +573,6 @@ public final class HConstants {
   public static int DEFAULT_HBASE_CLIENT_RETRIES_NUMBER = 10;
 
   /**
-   * Parameter name for maximum attempts, used to limit the number of times the
-   * client will try to obtain the proxy for a given region server.
-   */
-  public static String HBASE_CLIENT_RPC_MAXATTEMPTS = "hbase.client.rpc.maxattempts";
-
-  /**
-   * Default value of {@link #HBASE_CLIENT_RPC_MAXATTEMPTS}.
-   */
-  public static int DEFAULT_HBASE_CLIENT_RPC_MAXATTEMPTS = 1;
-
-  /**
    * Parameter name for client prefetch limit, used as the maximum number of regions
    * info that will be prefetched.
    */

Modified: hbase/branches/0.95/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java (original)
+++ hbase/branches/0.95/hbase-common/src/test/java/org/apache/hadoop/hbase/ClassFinder.java Fri May  3 03:58:33 2013
@@ -22,15 +22,14 @@ import java.io.File;
 import java.io.FileFilter;
 import java.io.FileInputStream;
 import java.io.IOException;
-import java.lang.reflect.Method;
-import java.lang.reflect.Modifier;
 import java.net.URL;
 import java.util.ArrayList;
 import java.util.Enumeration;
 import java.util.HashSet;
 import java.util.List;
 import java.util.Set;
-import java.util.jar.*;
+import java.util.jar.JarEntry;
+import java.util.jar.JarInputStream;
 import java.util.regex.Matcher;
 import java.util.regex.Pattern;
 

Modified: hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java (original)
+++ hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java Fri May  3 03:58:33 2013
@@ -23,15 +23,17 @@ import java.util.HashMap;
 import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.hbase.ClusterManager.ServiceType;
-import org.apache.hadoop.hbase.client.AdminProtocol;
-import org.apache.hadoop.hbase.client.ClientProtocol;
 import org.apache.hadoop.hbase.client.HBaseAdmin;
 import org.apache.hadoop.hbase.client.HConnection;
 import org.apache.hadoop.hbase.client.HConnectionManager;
 import org.apache.hadoop.hbase.exceptions.MasterNotRunningException;
 import org.apache.hadoop.hbase.exceptions.ZooKeeperConnectionException;
 import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
+import org.apache.hadoop.hbase.protobuf.generated.AdminProtos;
 import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ServerInfo;
+import org.apache.hadoop.hbase.protobuf.generated.ClientProtos;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos;
+import org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.hadoop.hbase.util.Threads;
 
@@ -86,12 +88,14 @@ public class DistributedHBaseCluster ext
   }
 
   @Override
-  public AdminProtocol getAdminProtocol(ServerName serverName) throws IOException {
+  public AdminProtos.AdminService.BlockingInterface getAdminProtocol(ServerName serverName)
+  throws IOException {
     return admin.getConnection().getAdmin(serverName);
   }
 
   @Override
-  public ClientProtocol getClientProtocol(ServerName serverName) throws IOException {
+  public ClientProtos.ClientService.BlockingInterface getClientProtocol(ServerName serverName)
+  throws IOException {
     return admin.getConnection().getClient(serverName);
   }
 
@@ -133,13 +137,15 @@ public class DistributedHBaseCluster ext
   }
 
   @Override
-  public MasterAdminProtocol getMasterAdmin() throws IOException {
+  public MasterAdminProtos.MasterAdminService.BlockingInterface getMasterAdmin()
+  throws IOException {
     HConnection conn = HConnectionManager.getConnection(conf);
     return conn.getMasterAdmin();
   }
 
   @Override
-  public MasterMonitorProtocol getMasterMonitor() throws IOException {
+  public MasterMonitorProtos.MasterMonitorService.BlockingInterface getMasterMonitor()
+  throws IOException {
     HConnection conn = HConnectionManager.getConnection(conf);
     return conn.getMasterMonitor();
   }
@@ -195,7 +201,8 @@ public class DistributedHBaseCluster ext
       return null;
     }
 
-    AdminProtocol client = connection.getAdmin(regionLoc.getServerName());
+    AdminProtos.AdminService.BlockingInterface client =
+      connection.getAdmin(regionLoc.getServerName());
     ServerInfo info = ProtobufUtil.getServerInfo(client);
     return ProtobufUtil.toServerName(info.getServerName());
   }

Modified: hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRebalanceAndKillServersTargeted.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRebalanceAndKillServersTargeted.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRebalanceAndKillServersTargeted.java (original)
+++ hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestRebalanceAndKillServersTargeted.java Fri May  3 03:58:33 2013
@@ -103,7 +103,7 @@ public class IntegrationTestRebalanceAnd
   @SuppressWarnings("unchecked")
   public void setUp() throws Exception {
     Configuration conf = HBaseConfiguration.create();
-    conf.set(HConnectionManager.RETRIES_BY_SERVER, "true");
+    conf.set(HConnectionManager.RETRIES_BY_SERVER_KEY, "true");
     super.setUp(NUM_SLAVES_BASE, conf);
 
     ChaosMonkey.Policy chaosPolicy = new ChaosMonkey.PeriodicRandomActionPolicy(

Modified: hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestsDriver.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestsDriver.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestsDriver.java (original)
+++ hbase/branches/0.95/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestsDriver.java Fri May  3 03:58:33 2013
@@ -19,7 +19,6 @@
 package org.apache.hadoop.hbase;
 
 import java.io.IOException;
-import java.util.List;
 import java.util.Set;
 import java.util.regex.Pattern;
 
@@ -113,5 +112,4 @@ public class IntegrationTestsDriver exte
 
     return result.wasSuccessful() ? 0 : 1;
   }
-
 }

Modified: hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterAdminProtos.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterAdminProtos.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterAdminProtos.java (original)
+++ hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterAdminProtos.java Fri May  3 03:58:33 2013
@@ -20646,6 +20646,11 @@ public final class MasterAdminProtos {
           org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneRequest request,
           com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse> done);
       
+      public abstract void isMasterRunning(
+          com.google.protobuf.RpcController controller,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request,
+          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse> done);
+      
     }
     
     public static com.google.protobuf.Service newReflectiveService(
@@ -20867,6 +20872,14 @@ public final class MasterAdminProtos {
           impl.isRestoreSnapshotDone(controller, request, done);
         }
         
+        @java.lang.Override
+        public  void isMasterRunning(
+            com.google.protobuf.RpcController controller,
+            org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request,
+            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse> done) {
+          impl.isMasterRunning(controller, request, done);
+        }
+        
       };
     }
     
@@ -20943,6 +20956,8 @@ public final class MasterAdminProtos {
               return impl.restoreSnapshot(controller, (org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.RestoreSnapshotRequest)request);
             case 26:
               return impl.isRestoreSnapshotDone(controller, (org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneRequest)request);
+            case 27:
+              return impl.isMasterRunning(controller, (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest)request);
             default:
               throw new java.lang.AssertionError("Can't get here.");
           }
@@ -21011,6 +21026,8 @@ public final class MasterAdminProtos {
               return org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.RestoreSnapshotRequest.getDefaultInstance();
             case 26:
               return org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneRequest.getDefaultInstance();
+            case 27:
+              return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest.getDefaultInstance();
             default:
               throw new java.lang.AssertionError("Can't get here.");
           }
@@ -21079,6 +21096,8 @@ public final class MasterAdminProtos {
               return org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.RestoreSnapshotResponse.getDefaultInstance();
             case 26:
               return org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse.getDefaultInstance();
+            case 27:
+              return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance();
             default:
               throw new java.lang.AssertionError("Can't get here.");
           }
@@ -21222,6 +21241,11 @@ public final class MasterAdminProtos {
         org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneRequest request,
         com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse> done);
     
+    public abstract void isMasterRunning(
+        com.google.protobuf.RpcController controller,
+        org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request,
+        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse> done);
+    
     public static final
         com.google.protobuf.Descriptors.ServiceDescriptor
         getDescriptor() {
@@ -21379,6 +21403,11 @@ public final class MasterAdminProtos {
             com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse>specializeCallback(
               done));
           return;
+        case 27:
+          this.isMasterRunning(controller, (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest)request,
+            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse>specializeCallback(
+              done));
+          return;
         default:
           throw new java.lang.AssertionError("Can't get here.");
       }
@@ -21447,6 +21476,8 @@ public final class MasterAdminProtos {
           return org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.RestoreSnapshotRequest.getDefaultInstance();
         case 26:
           return org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneRequest.getDefaultInstance();
+        case 27:
+          return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest.getDefaultInstance();
         default:
           throw new java.lang.AssertionError("Can't get here.");
       }
@@ -21515,6 +21546,8 @@ public final class MasterAdminProtos {
           return org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.RestoreSnapshotResponse.getDefaultInstance();
         case 26:
           return org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse.getDefaultInstance();
+        case 27:
+          return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance();
         default:
           throw new java.lang.AssertionError("Can't get here.");
       }
@@ -21940,6 +21973,21 @@ public final class MasterAdminProtos {
             org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse.class,
             org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse.getDefaultInstance()));
       }
+      
+      public  void isMasterRunning(
+          com.google.protobuf.RpcController controller,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request,
+          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse> done) {
+        channel.callMethod(
+          getDescriptor().getMethods().get(27),
+          controller,
+          request,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance(),
+          com.google.protobuf.RpcUtil.generalizeCallback(
+            done,
+            org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.class,
+            org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance()));
+      }
     }
     
     public static BlockingInterface newBlockingStub(
@@ -22082,6 +22130,11 @@ public final class MasterAdminProtos {
           com.google.protobuf.RpcController controller,
           org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneRequest request)
           throws com.google.protobuf.ServiceException;
+      
+      public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse isMasterRunning(
+          com.google.protobuf.RpcController controller,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request)
+          throws com.google.protobuf.ServiceException;
     }
     
     private static final class BlockingStub implements BlockingInterface {
@@ -22414,6 +22467,18 @@ public final class MasterAdminProtos {
           org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse.getDefaultInstance());
       }
       
+      
+      public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse isMasterRunning(
+          com.google.protobuf.RpcController controller,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request)
+          throws com.google.protobuf.ServiceException {
+        return (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse) channel.callBlockingMethod(
+          getDescriptor().getMethods().get(27),
+          controller,
+          request,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance());
+      }
+      
     }
   }
   
@@ -22686,114 +22751,117 @@ public final class MasterAdminProtos {
       descriptor;
   static {
     java.lang.String[] descriptorData = {
-      "\n\021MasterAdmin.proto\032\013hbase.proto\032\014Client" +
-      ".proto\"R\n\020AddColumnRequest\022\021\n\ttableName\030" +
-      "\001 \002(\014\022+\n\016columnFamilies\030\002 \002(\0132\023.ColumnFa" +
-      "milySchema\"\023\n\021AddColumnResponse\"<\n\023Delet" +
-      "eColumnRequest\022\021\n\ttableName\030\001 \002(\014\022\022\n\ncol" +
-      "umnName\030\002 \002(\014\"\026\n\024DeleteColumnResponse\"U\n" +
-      "\023ModifyColumnRequest\022\021\n\ttableName\030\001 \002(\014\022" +
-      "+\n\016columnFamilies\030\002 \002(\0132\023.ColumnFamilySc" +
-      "hema\"\026\n\024ModifyColumnResponse\"Z\n\021MoveRegi" +
-      "onRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpecif",
-      "ier\022#\n\016destServerName\030\002 \001(\0132\013.ServerName" +
-      "\"\024\n\022MoveRegionResponse\"~\n\035DispatchMergin" +
-      "gRegionsRequest\022!\n\007regionA\030\001 \002(\0132\020.Regio" +
-      "nSpecifier\022!\n\007regionB\030\002 \002(\0132\020.RegionSpec" +
-      "ifier\022\027\n\010forcible\030\003 \001(\010:\005false\" \n\036Dispat" +
-      "chMergingRegionsResponse\"7\n\023AssignRegion" +
-      "Request\022 \n\006region\030\001 \002(\0132\020.RegionSpecifie" +
-      "r\"\026\n\024AssignRegionResponse\"O\n\025UnassignReg" +
+      "\n\021MasterAdmin.proto\032\014Master.proto\032\013hbase" +
+      ".proto\032\014Client.proto\"R\n\020AddColumnRequest" +
+      "\022\021\n\ttableName\030\001 \002(\014\022+\n\016columnFamilies\030\002 " +
+      "\002(\0132\023.ColumnFamilySchema\"\023\n\021AddColumnRes" +
+      "ponse\"<\n\023DeleteColumnRequest\022\021\n\ttableNam" +
+      "e\030\001 \002(\014\022\022\n\ncolumnName\030\002 \002(\014\"\026\n\024DeleteCol" +
+      "umnResponse\"U\n\023ModifyColumnRequest\022\021\n\tta" +
+      "bleName\030\001 \002(\014\022+\n\016columnFamilies\030\002 \002(\0132\023." +
+      "ColumnFamilySchema\"\026\n\024ModifyColumnRespon" +
+      "se\"Z\n\021MoveRegionRequest\022 \n\006region\030\001 \002(\0132",
+      "\020.RegionSpecifier\022#\n\016destServerName\030\002 \001(" +
+      "\0132\013.ServerName\"\024\n\022MoveRegionResponse\"~\n\035" +
+      "DispatchMergingRegionsRequest\022!\n\007regionA" +
+      "\030\001 \002(\0132\020.RegionSpecifier\022!\n\007regionB\030\002 \002(" +
+      "\0132\020.RegionSpecifier\022\027\n\010forcible\030\003 \001(\010:\005f" +
+      "alse\" \n\036DispatchMergingRegionsResponse\"7" +
+      "\n\023AssignRegionRequest\022 \n\006region\030\001 \002(\0132\020." +
+      "RegionSpecifier\"\026\n\024AssignRegionResponse\"" +
+      "O\n\025UnassignRegionRequest\022 \n\006region\030\001 \002(\013" +
+      "2\020.RegionSpecifier\022\024\n\005force\030\002 \001(\010:\005false",
+      "\"\030\n\026UnassignRegionResponse\"8\n\024OfflineReg" +
       "ionRequest\022 \n\006region\030\001 \002(\0132\020.RegionSpeci" +
-      "fier\022\024\n\005force\030\002 \001(\010:\005false\"\030\n\026UnassignRe",
-      "gionResponse\"8\n\024OfflineRegionRequest\022 \n\006" +
-      "region\030\001 \002(\0132\020.RegionSpecifier\"\027\n\025Offlin" +
-      "eRegionResponse\"J\n\022CreateTableRequest\022!\n" +
-      "\013tableSchema\030\001 \002(\0132\014.TableSchema\022\021\n\tspli" +
-      "tKeys\030\002 \003(\014\"\025\n\023CreateTableResponse\"\'\n\022De" +
-      "leteTableRequest\022\021\n\ttableName\030\001 \002(\014\"\025\n\023D" +
-      "eleteTableResponse\"\'\n\022EnableTableRequest" +
-      "\022\021\n\ttableName\030\001 \002(\014\"\025\n\023EnableTableRespon" +
-      "se\"(\n\023DisableTableRequest\022\021\n\ttableName\030\001" +
-      " \002(\014\"\026\n\024DisableTableResponse\"J\n\022ModifyTa",
-      "bleRequest\022\021\n\ttableName\030\001 \002(\014\022!\n\013tableSc" +
-      "hema\030\002 \002(\0132\014.TableSchema\"\025\n\023ModifyTableR" +
-      "esponse\"\021\n\017ShutdownRequest\"\022\n\020ShutdownRe" +
-      "sponse\"\023\n\021StopMasterRequest\"\024\n\022StopMaste" +
-      "rResponse\"\020\n\016BalanceRequest\"&\n\017BalanceRe" +
-      "sponse\022\023\n\013balancerRan\030\001 \002(\010\"<\n\031SetBalanc" +
-      "erRunningRequest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchron" +
-      "ous\030\002 \001(\010\"6\n\032SetBalancerRunningResponse\022" +
-      "\030\n\020prevBalanceValue\030\001 \001(\010\"\024\n\022CatalogScan" +
-      "Request\")\n\023CatalogScanResponse\022\022\n\nscanRe",
-      "sult\030\001 \001(\005\"-\n\033EnableCatalogJanitorReques" +
-      "t\022\016\n\006enable\030\001 \002(\010\"1\n\034EnableCatalogJanito" +
-      "rResponse\022\021\n\tprevValue\030\001 \001(\010\" \n\036IsCatalo" +
-      "gJanitorEnabledRequest\"0\n\037IsCatalogJanit" +
-      "orEnabledResponse\022\r\n\005value\030\001 \002(\010\"=\n\023Take" +
-      "SnapshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Snap" +
-      "shotDescription\"/\n\024TakeSnapshotResponse\022" +
-      "\027\n\017expectedTimeout\030\001 \002(\003\"\025\n\023ListSnapshot" +
-      "Request\"?\n\024ListSnapshotResponse\022\'\n\tsnaps" +
-      "hots\030\001 \003(\0132\024.SnapshotDescription\"?\n\025Dele",
-      "teSnapshotRequest\022&\n\010snapshot\030\001 \002(\0132\024.Sn" +
-      "apshotDescription\"\030\n\026DeleteSnapshotRespo" +
-      "nse\"@\n\026RestoreSnapshotRequest\022&\n\010snapsho" +
-      "t\030\001 \002(\0132\024.SnapshotDescription\"\031\n\027Restore" +
-      "SnapshotResponse\"?\n\025IsSnapshotDoneReques" +
-      "t\022&\n\010snapshot\030\001 \001(\0132\024.SnapshotDescriptio" +
-      "n\"U\n\026IsSnapshotDoneResponse\022\023\n\004done\030\001 \001(" +
-      "\010:\005false\022&\n\010snapshot\030\002 \001(\0132\024.SnapshotDes" +
-      "cription\"F\n\034IsRestoreSnapshotDoneRequest" +
-      "\022&\n\010snapshot\030\001 \001(\0132\024.SnapshotDescription",
-      "\"3\n\035IsRestoreSnapshotDoneResponse\022\022\n\004don" +
-      "e\030\001 \001(\010:\004true2\377\r\n\022MasterAdminService\0222\n\t" +
-      "addColumn\022\021.AddColumnRequest\032\022.AddColumn" +
-      "Response\022;\n\014deleteColumn\022\024.DeleteColumnR" +
-      "equest\032\025.DeleteColumnResponse\022;\n\014modifyC" +
-      "olumn\022\024.ModifyColumnRequest\032\025.ModifyColu" +
-      "mnResponse\0225\n\nmoveRegion\022\022.MoveRegionReq" +
-      "uest\032\023.MoveRegionResponse\022Y\n\026dispatchMer" +
-      "gingRegions\022\036.DispatchMergingRegionsRequ" +
-      "est\032\037.DispatchMergingRegionsResponse\022;\n\014",
-      "assignRegion\022\024.AssignRegionRequest\032\025.Ass" +
-      "ignRegionResponse\022A\n\016unassignRegion\022\026.Un" +
-      "assignRegionRequest\032\027.UnassignRegionResp" +
-      "onse\022>\n\rofflineRegion\022\025.OfflineRegionReq" +
-      "uest\032\026.OfflineRegionResponse\0228\n\013deleteTa" +
-      "ble\022\023.DeleteTableRequest\032\024.DeleteTableRe" +
-      "sponse\0228\n\013enableTable\022\023.EnableTableReque" +
-      "st\032\024.EnableTableResponse\022;\n\014disableTable" +
-      "\022\024.DisableTableRequest\032\025.DisableTableRes" +
-      "ponse\0228\n\013modifyTable\022\023.ModifyTableReques",
-      "t\032\024.ModifyTableResponse\0228\n\013createTable\022\023" +
-      ".CreateTableRequest\032\024.CreateTableRespons" +
-      "e\022/\n\010shutdown\022\020.ShutdownRequest\032\021.Shutdo" +
-      "wnResponse\0225\n\nstopMaster\022\022.StopMasterReq" +
-      "uest\032\023.StopMasterResponse\022,\n\007balance\022\017.B" +
-      "alanceRequest\032\020.BalanceResponse\022M\n\022setBa" +
-      "lancerRunning\022\032.SetBalancerRunningReques" +
-      "t\032\033.SetBalancerRunningResponse\022;\n\016runCat" +
-      "alogScan\022\023.CatalogScanRequest\032\024.CatalogS" +
-      "canResponse\022S\n\024enableCatalogJanitor\022\034.En",
-      "ableCatalogJanitorRequest\032\035.EnableCatalo" +
-      "gJanitorResponse\022\\\n\027isCatalogJanitorEnab" +
-      "led\022\037.IsCatalogJanitorEnabledRequest\032 .I" +
-      "sCatalogJanitorEnabledResponse\022L\n\021execMa" +
-      "sterService\022\032.CoprocessorServiceRequest\032" +
-      "\033.CoprocessorServiceResponse\0227\n\010snapshot" +
-      "\022\024.TakeSnapshotRequest\032\025.TakeSnapshotRes" +
-      "ponse\022D\n\025getCompletedSnapshots\022\024.ListSna" +
-      "pshotRequest\032\025.ListSnapshotResponse\022A\n\016d" +
-      "eleteSnapshot\022\026.DeleteSnapshotRequest\032\027.",
-      "DeleteSnapshotResponse\022A\n\016isSnapshotDone" +
-      "\022\026.IsSnapshotDoneRequest\032\027.IsSnapshotDon" +
-      "eResponse\022D\n\017restoreSnapshot\022\027.RestoreSn" +
-      "apshotRequest\032\030.RestoreSnapshotResponse\022" +
-      "V\n\025isRestoreSnapshotDone\022\035.IsRestoreSnap" +
-      "shotDoneRequest\032\036.IsRestoreSnapshotDoneR" +
-      "esponseBG\n*org.apache.hadoop.hbase.proto" +
-      "buf.generatedB\021MasterAdminProtosH\001\210\001\001\240\001\001"
+      "fier\"\027\n\025OfflineRegionResponse\"J\n\022CreateT" +
+      "ableRequest\022!\n\013tableSchema\030\001 \002(\0132\014.Table" +
+      "Schema\022\021\n\tsplitKeys\030\002 \003(\014\"\025\n\023CreateTable" +
+      "Response\"\'\n\022DeleteTableRequest\022\021\n\ttableN" +
+      "ame\030\001 \002(\014\"\025\n\023DeleteTableResponse\"\'\n\022Enab" +
+      "leTableRequest\022\021\n\ttableName\030\001 \002(\014\"\025\n\023Ena" +
+      "bleTableResponse\"(\n\023DisableTableRequest\022" +
+      "\021\n\ttableName\030\001 \002(\014\"\026\n\024DisableTableRespon",
+      "se\"J\n\022ModifyTableRequest\022\021\n\ttableName\030\001 " +
+      "\002(\014\022!\n\013tableSchema\030\002 \002(\0132\014.TableSchema\"\025" +
+      "\n\023ModifyTableResponse\"\021\n\017ShutdownRequest" +
+      "\"\022\n\020ShutdownResponse\"\023\n\021StopMasterReques" +
+      "t\"\024\n\022StopMasterResponse\"\020\n\016BalanceReques" +
+      "t\"&\n\017BalanceResponse\022\023\n\013balancerRan\030\001 \002(" +
+      "\010\"<\n\031SetBalancerRunningRequest\022\n\n\002on\030\001 \002" +
+      "(\010\022\023\n\013synchronous\030\002 \001(\010\"6\n\032SetBalancerRu" +
+      "nningResponse\022\030\n\020prevBalanceValue\030\001 \001(\010\"" +
+      "\024\n\022CatalogScanRequest\")\n\023CatalogScanResp",
+      "onse\022\022\n\nscanResult\030\001 \001(\005\"-\n\033EnableCatalo" +
+      "gJanitorRequest\022\016\n\006enable\030\001 \002(\010\"1\n\034Enabl" +
+      "eCatalogJanitorResponse\022\021\n\tprevValue\030\001 \001" +
+      "(\010\" \n\036IsCatalogJanitorEnabledRequest\"0\n\037" +
+      "IsCatalogJanitorEnabledResponse\022\r\n\005value" +
+      "\030\001 \002(\010\"=\n\023TakeSnapshotRequest\022&\n\010snapsho" +
+      "t\030\001 \002(\0132\024.SnapshotDescription\"/\n\024TakeSna" +
+      "pshotResponse\022\027\n\017expectedTimeout\030\001 \002(\003\"\025" +
+      "\n\023ListSnapshotRequest\"?\n\024ListSnapshotRes" +
+      "ponse\022\'\n\tsnapshots\030\001 \003(\0132\024.SnapshotDescr",
+      "iption\"?\n\025DeleteSnapshotRequest\022&\n\010snaps" +
+      "hot\030\001 \002(\0132\024.SnapshotDescription\"\030\n\026Delet" +
+      "eSnapshotResponse\"@\n\026RestoreSnapshotRequ" +
+      "est\022&\n\010snapshot\030\001 \002(\0132\024.SnapshotDescript" +
+      "ion\"\031\n\027RestoreSnapshotResponse\"?\n\025IsSnap" +
+      "shotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snap" +
+      "shotDescription\"U\n\026IsSnapshotDoneRespons" +
+      "e\022\023\n\004done\030\001 \001(\010:\005false\022&\n\010snapshot\030\002 \001(\013" +
+      "2\024.SnapshotDescription\"F\n\034IsRestoreSnaps" +
+      "hotDoneRequest\022&\n\010snapshot\030\001 \001(\0132\024.Snaps",
+      "hotDescription\"3\n\035IsRestoreSnapshotDoneR" +
+      "esponse\022\022\n\004done\030\001 \001(\010:\004true2\305\016\n\022MasterAd" +
+      "minService\0222\n\taddColumn\022\021.AddColumnReque" +
+      "st\032\022.AddColumnResponse\022;\n\014deleteColumn\022\024" +
+      ".DeleteColumnRequest\032\025.DeleteColumnRespo" +
+      "nse\022;\n\014modifyColumn\022\024.ModifyColumnReques" +
+      "t\032\025.ModifyColumnResponse\0225\n\nmoveRegion\022\022" +
+      ".MoveRegionRequest\032\023.MoveRegionResponse\022" +
+      "Y\n\026dispatchMergingRegions\022\036.DispatchMerg" +
+      "ingRegionsRequest\032\037.DispatchMergingRegio",
+      "nsResponse\022;\n\014assignRegion\022\024.AssignRegio" +
+      "nRequest\032\025.AssignRegionResponse\022A\n\016unass" +
+      "ignRegion\022\026.UnassignRegionRequest\032\027.Unas" +
+      "signRegionResponse\022>\n\rofflineRegion\022\025.Of" +
+      "flineRegionRequest\032\026.OfflineRegionRespon" +
+      "se\0228\n\013deleteTable\022\023.DeleteTableRequest\032\024" +
+      ".DeleteTableResponse\0228\n\013enableTable\022\023.En" +
+      "ableTableRequest\032\024.EnableTableResponse\022;" +
+      "\n\014disableTable\022\024.DisableTableRequest\032\025.D" +
+      "isableTableResponse\0228\n\013modifyTable\022\023.Mod",
+      "ifyTableRequest\032\024.ModifyTableResponse\0228\n" +
+      "\013createTable\022\023.CreateTableRequest\032\024.Crea" +
+      "teTableResponse\022/\n\010shutdown\022\020.ShutdownRe" +
+      "quest\032\021.ShutdownResponse\0225\n\nstopMaster\022\022" +
+      ".StopMasterRequest\032\023.StopMasterResponse\022" +
+      ",\n\007balance\022\017.BalanceRequest\032\020.BalanceRes" +
+      "ponse\022M\n\022setBalancerRunning\022\032.SetBalance" +
+      "rRunningRequest\032\033.SetBalancerRunningResp" +
+      "onse\022;\n\016runCatalogScan\022\023.CatalogScanRequ" +
+      "est\032\024.CatalogScanResponse\022S\n\024enableCatal",
+      "ogJanitor\022\034.EnableCatalogJanitorRequest\032" +
+      "\035.EnableCatalogJanitorResponse\022\\\n\027isCata" +
+      "logJanitorEnabled\022\037.IsCatalogJanitorEnab" +
+      "ledRequest\032 .IsCatalogJanitorEnabledResp" +
+      "onse\022L\n\021execMasterService\022\032.CoprocessorS" +
+      "erviceRequest\032\033.CoprocessorServiceRespon" +
+      "se\0227\n\010snapshot\022\024.TakeSnapshotRequest\032\025.T" +
+      "akeSnapshotResponse\022D\n\025getCompletedSnaps" +
+      "hots\022\024.ListSnapshotRequest\032\025.ListSnapsho" +
+      "tResponse\022A\n\016deleteSnapshot\022\026.DeleteSnap",
+      "shotRequest\032\027.DeleteSnapshotResponse\022A\n\016" +
+      "isSnapshotDone\022\026.IsSnapshotDoneRequest\032\027" +
+      ".IsSnapshotDoneResponse\022D\n\017restoreSnapsh" +
+      "ot\022\027.RestoreSnapshotRequest\032\030.RestoreSna" +
+      "pshotResponse\022V\n\025isRestoreSnapshotDone\022\035" +
+      ".IsRestoreSnapshotDoneRequest\032\036.IsRestor" +
+      "eSnapshotDoneResponse\022D\n\017isMasterRunning" +
+      "\022\027.IsMasterRunningRequest\032\030.IsMasterRunn" +
+      "ingResponseBG\n*org.apache.hadoop.hbase.p" +
+      "rotobuf.generatedB\021MasterAdminProtosH\001\210\001",
+      "\001\240\001\001"
     };
     com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
       new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -23222,6 +23290,7 @@ public final class MasterAdminProtos {
     com.google.protobuf.Descriptors.FileDescriptor
       .internalBuildGeneratedFileFrom(descriptorData,
         new com.google.protobuf.Descriptors.FileDescriptor[] {
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.getDescriptor(),
           org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
           org.apache.hadoop.hbase.protobuf.generated.ClientProtos.getDescriptor(),
         }, assigner);

Modified: hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterMonitorProtos.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterMonitorProtos.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterMonitorProtos.java (original)
+++ hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/MasterMonitorProtos.java Fri May  3 03:58:33 2013
@@ -2632,6 +2632,11 @@ public final class MasterMonitorProtos {
           org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusRequest request,
           com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusResponse> done);
       
+      public abstract void isMasterRunning(
+          com.google.protobuf.RpcController controller,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request,
+          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse> done);
+      
     }
     
     public static com.google.protobuf.Service newReflectiveService(
@@ -2661,6 +2666,14 @@ public final class MasterMonitorProtos {
           impl.getClusterStatus(controller, request, done);
         }
         
+        @java.lang.Override
+        public  void isMasterRunning(
+            com.google.protobuf.RpcController controller,
+            org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request,
+            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse> done) {
+          impl.isMasterRunning(controller, request, done);
+        }
+        
       };
     }
     
@@ -2689,6 +2702,8 @@ public final class MasterMonitorProtos {
               return impl.getTableDescriptors(controller, (org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetTableDescriptorsRequest)request);
             case 2:
               return impl.getClusterStatus(controller, (org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusRequest)request);
+            case 3:
+              return impl.isMasterRunning(controller, (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest)request);
             default:
               throw new java.lang.AssertionError("Can't get here.");
           }
@@ -2709,6 +2724,8 @@ public final class MasterMonitorProtos {
               return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetTableDescriptorsRequest.getDefaultInstance();
             case 2:
               return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusRequest.getDefaultInstance();
+            case 3:
+              return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest.getDefaultInstance();
             default:
               throw new java.lang.AssertionError("Can't get here.");
           }
@@ -2729,6 +2746,8 @@ public final class MasterMonitorProtos {
               return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetTableDescriptorsResponse.getDefaultInstance();
             case 2:
               return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusResponse.getDefaultInstance();
+            case 3:
+              return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance();
             default:
               throw new java.lang.AssertionError("Can't get here.");
           }
@@ -2752,6 +2771,11 @@ public final class MasterMonitorProtos {
         org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusRequest request,
         com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusResponse> done);
     
+    public abstract void isMasterRunning(
+        com.google.protobuf.RpcController controller,
+        org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request,
+        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse> done);
+    
     public static final
         com.google.protobuf.Descriptors.ServiceDescriptor
         getDescriptor() {
@@ -2789,6 +2813,11 @@ public final class MasterMonitorProtos {
             com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusResponse>specializeCallback(
               done));
           return;
+        case 3:
+          this.isMasterRunning(controller, (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest)request,
+            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse>specializeCallback(
+              done));
+          return;
         default:
           throw new java.lang.AssertionError("Can't get here.");
       }
@@ -2809,6 +2838,8 @@ public final class MasterMonitorProtos {
           return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetTableDescriptorsRequest.getDefaultInstance();
         case 2:
           return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusRequest.getDefaultInstance();
+        case 3:
+          return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest.getDefaultInstance();
         default:
           throw new java.lang.AssertionError("Can't get here.");
       }
@@ -2829,6 +2860,8 @@ public final class MasterMonitorProtos {
           return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetTableDescriptorsResponse.getDefaultInstance();
         case 2:
           return org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusResponse.getDefaultInstance();
+        case 3:
+          return org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance();
         default:
           throw new java.lang.AssertionError("Can't get here.");
       }
@@ -2894,6 +2927,21 @@ public final class MasterMonitorProtos {
             org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusResponse.class,
             org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusResponse.getDefaultInstance()));
       }
+      
+      public  void isMasterRunning(
+          com.google.protobuf.RpcController controller,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request,
+          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse> done) {
+        channel.callMethod(
+          getDescriptor().getMethods().get(3),
+          controller,
+          request,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance(),
+          com.google.protobuf.RpcUtil.generalizeCallback(
+            done,
+            org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.class,
+            org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance()));
+      }
     }
     
     public static BlockingInterface newBlockingStub(
@@ -2916,6 +2964,11 @@ public final class MasterMonitorProtos {
           com.google.protobuf.RpcController controller,
           org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusRequest request)
           throws com.google.protobuf.ServiceException;
+      
+      public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse isMasterRunning(
+          com.google.protobuf.RpcController controller,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request)
+          throws com.google.protobuf.ServiceException;
     }
     
     private static final class BlockingStub implements BlockingInterface {
@@ -2960,6 +3013,18 @@ public final class MasterMonitorProtos {
           org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos.GetClusterStatusResponse.getDefaultInstance());
       }
       
+      
+      public org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse isMasterRunning(
+          com.google.protobuf.RpcController controller,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningRequest request)
+          throws com.google.protobuf.ServiceException {
+        return (org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse) channel.callBlockingMethod(
+          getDescriptor().getMethods().get(3),
+          controller,
+          request,
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance());
+      }
+      
     }
   }
   
@@ -3002,25 +3067,27 @@ public final class MasterMonitorProtos {
       descriptor;
   static {
     java.lang.String[] descriptorData = {
-      "\n\023MasterMonitor.proto\032\013hbase.proto\032\023Clus" +
-      "terStatus.proto\"0\n\033GetSchemaAlterStatusR" +
-      "equest\022\021\n\ttableName\030\001 \002(\014\"P\n\034GetSchemaAl" +
-      "terStatusResponse\022\032\n\022yetToUpdateRegions\030" +
-      "\001 \001(\r\022\024\n\014totalRegions\030\002 \001(\r\"0\n\032GetTableD" +
-      "escriptorsRequest\022\022\n\ntableNames\030\001 \003(\t\"@\n" +
-      "\033GetTableDescriptorsResponse\022!\n\013tableSch" +
-      "ema\030\001 \003(\0132\014.TableSchema\"\031\n\027GetClusterSta" +
-      "tusRequest\"A\n\030GetClusterStatusResponse\022%" +
-      "\n\rclusterStatus\030\001 \002(\0132\016.ClusterStatus2\206\002",
-      "\n\024MasterMonitorService\022S\n\024getSchemaAlter" +
-      "Status\022\034.GetSchemaAlterStatusRequest\032\035.G" +
-      "etSchemaAlterStatusResponse\022P\n\023getTableD" +
-      "escriptors\022\033.GetTableDescriptorsRequest\032" +
-      "\034.GetTableDescriptorsResponse\022G\n\020getClus" +
-      "terStatus\022\030.GetClusterStatusRequest\032\031.Ge" +
-      "tClusterStatusResponseBI\n*org.apache.had" +
-      "oop.hbase.protobuf.generatedB\023MasterMoni" +
-      "torProtosH\001\210\001\001\240\001\001"
+      "\n\023MasterMonitor.proto\032\014Master.proto\032\013hba" +
+      "se.proto\032\023ClusterStatus.proto\"0\n\033GetSche" +
+      "maAlterStatusRequest\022\021\n\ttableName\030\001 \002(\014\"" +
+      "P\n\034GetSchemaAlterStatusResponse\022\032\n\022yetTo" +
+      "UpdateRegions\030\001 \001(\r\022\024\n\014totalRegions\030\002 \001(" +
+      "\r\"0\n\032GetTableDescriptorsRequest\022\022\n\ntable" +
+      "Names\030\001 \003(\t\"@\n\033GetTableDescriptorsRespon" +
+      "se\022!\n\013tableSchema\030\001 \003(\0132\014.TableSchema\"\031\n" +
+      "\027GetClusterStatusRequest\"A\n\030GetClusterSt" +
+      "atusResponse\022%\n\rclusterStatus\030\001 \002(\0132\016.Cl",
+      "usterStatus2\314\002\n\024MasterMonitorService\022S\n\024" +
+      "getSchemaAlterStatus\022\034.GetSchemaAlterSta" +
+      "tusRequest\032\035.GetSchemaAlterStatusRespons" +
+      "e\022P\n\023getTableDescriptors\022\033.GetTableDescr" +
+      "iptorsRequest\032\034.GetTableDescriptorsRespo" +
+      "nse\022G\n\020getClusterStatus\022\030.GetClusterStat" +
+      "usRequest\032\031.GetClusterStatusResponse\022D\n\017" +
+      "isMasterRunning\022\027.IsMasterRunningRequest" +
+      "\032\030.IsMasterRunningResponseBI\n*org.apache" +
+      ".hadoop.hbase.protobuf.generatedB\023Master",
+      "MonitorProtosH\001\210\001\001\240\001\001"
     };
     com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
       new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -3081,6 +3148,7 @@ public final class MasterMonitorProtos {
     com.google.protobuf.Descriptors.FileDescriptor
       .internalBuildGeneratedFileFrom(descriptorData,
         new com.google.protobuf.Descriptors.FileDescriptor[] {
+          org.apache.hadoop.hbase.protobuf.generated.MasterProtos.getDescriptor(),
           org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(),
           org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.getDescriptor(),
         }, assigner);

Modified: hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java (original)
+++ hbase/branches/0.95/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/RPCProtos.java Fri May  3 03:58:33 2013
@@ -535,9 +535,9 @@ public final class RPCProtos {
     org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation getUserInfo();
     org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformationOrBuilder getUserInfoOrBuilder();
     
-    // optional string protocol = 2 [default = "org.apache.hadoop.hbase.client.ClientProtocol"];
-    boolean hasProtocol();
-    String getProtocol();
+    // optional string serviceName = 2;
+    boolean hasServiceName();
+    String getServiceName();
     
     // optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];
     boolean hasCellBlockCodecClass();
@@ -589,14 +589,14 @@ public final class RPCProtos {
       return userInfo_;
     }
     
-    // optional string protocol = 2 [default = "org.apache.hadoop.hbase.client.ClientProtocol"];
-    public static final int PROTOCOL_FIELD_NUMBER = 2;
-    private java.lang.Object protocol_;
-    public boolean hasProtocol() {
+    // optional string serviceName = 2;
+    public static final int SERVICENAME_FIELD_NUMBER = 2;
+    private java.lang.Object serviceName_;
+    public boolean hasServiceName() {
       return ((bitField0_ & 0x00000002) == 0x00000002);
     }
-    public String getProtocol() {
-      java.lang.Object ref = protocol_;
+    public String getServiceName() {
+      java.lang.Object ref = serviceName_;
       if (ref instanceof String) {
         return (String) ref;
       } else {
@@ -604,17 +604,17 @@ public final class RPCProtos {
             (com.google.protobuf.ByteString) ref;
         String s = bs.toStringUtf8();
         if (com.google.protobuf.Internal.isValidUtf8(bs)) {
-          protocol_ = s;
+          serviceName_ = s;
         }
         return s;
       }
     }
-    private com.google.protobuf.ByteString getProtocolBytes() {
-      java.lang.Object ref = protocol_;
+    private com.google.protobuf.ByteString getServiceNameBytes() {
+      java.lang.Object ref = serviceName_;
       if (ref instanceof String) {
         com.google.protobuf.ByteString b = 
             com.google.protobuf.ByteString.copyFromUtf8((String) ref);
-        protocol_ = b;
+        serviceName_ = b;
         return b;
       } else {
         return (com.google.protobuf.ByteString) ref;
@@ -687,7 +687,7 @@ public final class RPCProtos {
     
     private void initFields() {
       userInfo_ = org.apache.hadoop.hbase.protobuf.generated.RPCProtos.UserInformation.getDefaultInstance();
-      protocol_ = "org.apache.hadoop.hbase.client.ClientProtocol";
+      serviceName_ = "";
       cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec";
       cellBlockCompressorClass_ = "";
     }
@@ -713,7 +713,7 @@ public final class RPCProtos {
         output.writeMessage(1, userInfo_);
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeBytes(2, getProtocolBytes());
+        output.writeBytes(2, getServiceNameBytes());
       }
       if (((bitField0_ & 0x00000004) == 0x00000004)) {
         output.writeBytes(3, getCellBlockCodecClassBytes());
@@ -736,7 +736,7 @@ public final class RPCProtos {
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeBytesSize(2, getProtocolBytes());
+          .computeBytesSize(2, getServiceNameBytes());
       }
       if (((bitField0_ & 0x00000004) == 0x00000004)) {
         size += com.google.protobuf.CodedOutputStream
@@ -774,10 +774,10 @@ public final class RPCProtos {
         result = result && getUserInfo()
             .equals(other.getUserInfo());
       }
-      result = result && (hasProtocol() == other.hasProtocol());
-      if (hasProtocol()) {
-        result = result && getProtocol()
-            .equals(other.getProtocol());
+      result = result && (hasServiceName() == other.hasServiceName());
+      if (hasServiceName()) {
+        result = result && getServiceName()
+            .equals(other.getServiceName());
       }
       result = result && (hasCellBlockCodecClass() == other.hasCellBlockCodecClass());
       if (hasCellBlockCodecClass()) {
@@ -802,9 +802,9 @@ public final class RPCProtos {
         hash = (37 * hash) + USERINFO_FIELD_NUMBER;
         hash = (53 * hash) + getUserInfo().hashCode();
       }
-      if (hasProtocol()) {
-        hash = (37 * hash) + PROTOCOL_FIELD_NUMBER;
-        hash = (53 * hash) + getProtocol().hashCode();
+      if (hasServiceName()) {
+        hash = (37 * hash) + SERVICENAME_FIELD_NUMBER;
+        hash = (53 * hash) + getServiceName().hashCode();
       }
       if (hasCellBlockCodecClass()) {
         hash = (37 * hash) + CELLBLOCKCODECCLASS_FIELD_NUMBER;
@@ -937,7 +937,7 @@ public final class RPCProtos {
           userInfoBuilder_.clear();
         }
         bitField0_ = (bitField0_ & ~0x00000001);
-        protocol_ = "org.apache.hadoop.hbase.client.ClientProtocol";
+        serviceName_ = "";
         bitField0_ = (bitField0_ & ~0x00000002);
         cellBlockCodecClass_ = "org.apache.hadoop.hbase.codec.KeyValueCodec";
         bitField0_ = (bitField0_ & ~0x00000004);
@@ -992,7 +992,7 @@ public final class RPCProtos {
         if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
           to_bitField0_ |= 0x00000002;
         }
-        result.protocol_ = protocol_;
+        result.serviceName_ = serviceName_;
         if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
           to_bitField0_ |= 0x00000004;
         }
@@ -1020,8 +1020,8 @@ public final class RPCProtos {
         if (other.hasUserInfo()) {
           mergeUserInfo(other.getUserInfo());
         }
-        if (other.hasProtocol()) {
-          setProtocol(other.getProtocol());
+        if (other.hasServiceName()) {
+          setServiceName(other.getServiceName());
         }
         if (other.hasCellBlockCodecClass()) {
           setCellBlockCodecClass(other.getCellBlockCodecClass());
@@ -1077,7 +1077,7 @@ public final class RPCProtos {
             }
             case 18: {
               bitField0_ |= 0x00000002;
-              protocol_ = input.readBytes();
+              serviceName_ = input.readBytes();
               break;
             }
             case 26: {
@@ -1186,39 +1186,39 @@ public final class RPCProtos {
         return userInfoBuilder_;
       }
       
-      // optional string protocol = 2 [default = "org.apache.hadoop.hbase.client.ClientProtocol"];
-      private java.lang.Object protocol_ = "org.apache.hadoop.hbase.client.ClientProtocol";
-      public boolean hasProtocol() {
+      // optional string serviceName = 2;
+      private java.lang.Object serviceName_ = "";
+      public boolean hasServiceName() {
         return ((bitField0_ & 0x00000002) == 0x00000002);
       }
-      public String getProtocol() {
-        java.lang.Object ref = protocol_;
+      public String getServiceName() {
+        java.lang.Object ref = serviceName_;
         if (!(ref instanceof String)) {
           String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
-          protocol_ = s;
+          serviceName_ = s;
           return s;
         } else {
           return (String) ref;
         }
       }
-      public Builder setProtocol(String value) {
+      public Builder setServiceName(String value) {
         if (value == null) {
     throw new NullPointerException();
   }
   bitField0_ |= 0x00000002;
-        protocol_ = value;
+        serviceName_ = value;
         onChanged();
         return this;
       }
-      public Builder clearProtocol() {
+      public Builder clearServiceName() {
         bitField0_ = (bitField0_ & ~0x00000002);
-        protocol_ = getDefaultInstance().getProtocol();
+        serviceName_ = getDefaultInstance().getServiceName();
         onChanged();
         return this;
       }
-      void setProtocol(com.google.protobuf.ByteString value) {
+      void setServiceName(com.google.protobuf.ByteString value) {
         bitField0_ |= 0x00000002;
-        protocol_ = value;
+        serviceName_ = value;
         onChanged();
       }
       
@@ -3982,25 +3982,23 @@ public final class RPCProtos {
     java.lang.String[] descriptorData = {
       "\n\tRPC.proto\032\rTracing.proto\032\013hbase.proto\"" +
       ":\n\017UserInformation\022\025\n\reffectiveUser\030\001 \002(" +
-      "\t\022\020\n\010realUser\030\002 \001(\t\"\343\001\n\020ConnectionHeader" +
-      "\022\"\n\010userInfo\030\001 \001(\0132\020.UserInformation\022?\n\010" +
-      "protocol\030\002 \001(\t:-org.apache.hadoop.hbase." +
-      "client.ClientProtocol\022H\n\023cellBlockCodecC" +
-      "lass\030\003 \001(\t:+org.apache.hadoop.hbase.code" +
-      "c.KeyValueCodec\022 \n\030cellBlockCompressorCl" +
-      "ass\030\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006length\030\001 \001" +
-      "(\r\"w\n\021ExceptionResponse\022\032\n\022exceptionClas",
-      "sName\030\001 \001(\t\022\022\n\nstackTrace\030\002 \001(\t\022\020\n\010hostn" +
-      "ame\030\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\022\n\ndoNotRetry\030\005 " +
-      "\001(\010\"\216\001\n\rRequestHeader\022\016\n\006callId\030\001 \001(\r\022\034\n" +
-      "\ttraceInfo\030\002 \001(\0132\t.RPCTInfo\022\022\n\nmethodNam" +
-      "e\030\003 \001(\t\022\024\n\014requestParam\030\004 \001(\010\022%\n\rcellBlo" +
-      "ckMeta\030\005 \001(\0132\016.CellBlockMeta\"n\n\016Response" +
-      "Header\022\016\n\006callId\030\001 \001(\r\022%\n\texception\030\002 \001(" +
-      "\0132\022.ExceptionResponse\022%\n\rcellBlockMeta\030\003" +
-      " \001(\0132\016.CellBlockMetaB<\n*org.apache.hadoo" +
-      "p.hbase.protobuf.generatedB\tRPCProtosH\001\240",
-      "\001\001"
+      "\t\022\020\n\010realUser\030\002 \001(\t\"\267\001\n\020ConnectionHeader" +
+      "\022\"\n\010userInfo\030\001 \001(\0132\020.UserInformation\022\023\n\013" +
+      "serviceName\030\002 \001(\t\022H\n\023cellBlockCodecClass" +
+      "\030\003 \001(\t:+org.apache.hadoop.hbase.codec.Ke" +
+      "yValueCodec\022 \n\030cellBlockCompressorClass\030" +
+      "\004 \001(\t\"\037\n\rCellBlockMeta\022\016\n\006length\030\001 \001(\r\"w" +
+      "\n\021ExceptionResponse\022\032\n\022exceptionClassNam" +
+      "e\030\001 \001(\t\022\022\n\nstackTrace\030\002 \001(\t\022\020\n\010hostname\030",
+      "\003 \001(\t\022\014\n\004port\030\004 \001(\005\022\022\n\ndoNotRetry\030\005 \001(\010\"" +
+      "\216\001\n\rRequestHeader\022\016\n\006callId\030\001 \001(\r\022\034\n\ttra" +
+      "ceInfo\030\002 \001(\0132\t.RPCTInfo\022\022\n\nmethodName\030\003 " +
+      "\001(\t\022\024\n\014requestParam\030\004 \001(\010\022%\n\rcellBlockMe" +
+      "ta\030\005 \001(\0132\016.CellBlockMeta\"n\n\016ResponseHead" +
+      "er\022\016\n\006callId\030\001 \001(\r\022%\n\texception\030\002 \001(\0132\022." +
+      "ExceptionResponse\022%\n\rcellBlockMeta\030\003 \001(\013" +
+      "2\016.CellBlockMetaB<\n*org.apache.hadoop.hb" +
+      "ase.protobuf.generatedB\tRPCProtosH\001\240\001\001"
     };
     com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
       new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
@@ -4020,7 +4018,7 @@ public final class RPCProtos {
           internal_static_ConnectionHeader_fieldAccessorTable = new
             com.google.protobuf.GeneratedMessage.FieldAccessorTable(
               internal_static_ConnectionHeader_descriptor,
-              new java.lang.String[] { "UserInfo", "Protocol", "CellBlockCodecClass", "CellBlockCompressorClass", },
+              new java.lang.String[] { "UserInfo", "ServiceName", "CellBlockCodecClass", "CellBlockCompressorClass", },
               org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ConnectionHeader.class,
               org.apache.hadoop.hbase.protobuf.generated.RPCProtos.ConnectionHeader.Builder.class);
           internal_static_CellBlockMeta_descriptor =

Modified: hbase/branches/0.95/hbase-protocol/src/main/protobuf/MasterAdmin.proto
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-protocol/src/main/protobuf/MasterAdmin.proto?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-protocol/src/main/protobuf/MasterAdmin.proto (original)
+++ hbase/branches/0.95/hbase-protocol/src/main/protobuf/MasterAdmin.proto Fri May  3 03:58:33 2013
@@ -18,6 +18,8 @@
 
 // This file contains protocol buffers that are used for MasterAdminProtocol.
 
+import "Master.proto";
+
 option java_package = "org.apache.hadoop.hbase.protobuf.generated";
 option java_outer_classname = "MasterAdminProtos";
 option java_generic_services = true;
@@ -354,7 +356,7 @@ service MasterAdminService {
 
   /**
    * List completed snapshots.
-   * @return a list of snapshot descriptors for completed snapshots
+   * Returns a list of snapshot descriptors for completed snapshots
    */
   rpc getCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
 
@@ -379,4 +381,7 @@ service MasterAdminService {
    * Determine if the snapshot restore is done yet.
    */
   rpc isRestoreSnapshotDone(IsRestoreSnapshotDoneRequest) returns(IsRestoreSnapshotDoneResponse);
+
+  /** return true if master is available */
+  rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
 }

Modified: hbase/branches/0.95/hbase-protocol/src/main/protobuf/MasterMonitor.proto
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-protocol/src/main/protobuf/MasterMonitor.proto?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-protocol/src/main/protobuf/MasterMonitor.proto (original)
+++ hbase/branches/0.95/hbase-protocol/src/main/protobuf/MasterMonitor.proto Fri May  3 03:58:33 2013
@@ -17,6 +17,7 @@
  */
 
 // This file contains protocol buffers that are used for MasterMonitorProtocol.
+import "Master.proto";
 
 option java_package = "org.apache.hadoop.hbase.protobuf.generated";
 option java_outer_classname = "MasterMonitorProtos";
@@ -63,4 +64,7 @@ service MasterMonitorService {
   /** Return cluster status. */
   rpc getClusterStatus(GetClusterStatusRequest)
     returns(GetClusterStatusResponse);
+
+  /** return true if master is available */
+  rpc isMasterRunning(IsMasterRunningRequest) returns(IsMasterRunningResponse);
 }

Modified: hbase/branches/0.95/hbase-protocol/src/main/protobuf/RPC.proto
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-protocol/src/main/protobuf/RPC.proto?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-protocol/src/main/protobuf/RPC.proto (original)
+++ hbase/branches/0.95/hbase-protocol/src/main/protobuf/RPC.proto Fri May  3 03:58:33 2013
@@ -79,7 +79,7 @@ message UserInformation {
 // This is sent on connection setup after the connection preamble is sent.
 message ConnectionHeader {
   optional UserInformation userInfo = 1;
-  optional string protocol = 2 [default = "org.apache.hadoop.hbase.client.ClientProtocol"];
+  optional string serviceName = 2;
   // Cell block codec we will use sending over optional cell blocks.  Server throws exception
   // if cannot deal.
   optional string cellBlockCodecClass = 3 [default = "org.apache.hadoop.hbase.codec.KeyValueCodec"];

Modified: hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ZNodeClearer.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ZNodeClearer.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ZNodeClearer.java (original)
+++ hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ZNodeClearer.java Fri May  3 03:58:33 2013
@@ -53,10 +53,9 @@ public class ZNodeClearer {
    */
   public static void writeMyEphemeralNodeOnDisk(String fileContent) {
     String fileName = ZNodeClearer.getMyEphemeralNodeFileName();
-
     if (fileName == null) {
-      LOG.warn("No filename given to save the znode used, it won't be saved " +
-          "(Environment variable HBASE_ZNODE_FILE is not set).");
+      LOG.warn("Environment variable HBASE_ZNODE_FILE not set; znodes will not be cleared " +
+        "on crash by start scripts (Longer MTTR!)");
       return;
     }
 

Modified: hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/fs/HFileSystem.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/fs/HFileSystem.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/fs/HFileSystem.java (original)
+++ hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/fs/HFileSystem.java Fri May  3 03:58:33 2013
@@ -196,8 +196,6 @@ public class HFileSystem extends FilterF
    * @return true if the interceptor was added, false otherwise.
    */
   static boolean addLocationsOrderInterceptor(Configuration conf, final ReorderBlocks lrb) {
-    LOG.debug("Starting addLocationsOrderInterceptor with class " + lrb.getClass());
-
     if (!conf.getBoolean("hbase.filesystem.reorder.blocks", true)) {  // activated by default
       LOG.debug("addLocationsOrderInterceptor configured to false");
       return false;
@@ -212,8 +210,8 @@ public class HFileSystem extends FilterF
     }
 
     if (!(fs instanceof DistributedFileSystem)) {
-      LOG.warn("The file system is not a DistributedFileSystem." +
-          "Not adding block location reordering");
+      LOG.debug("The file system is not a DistributedFileSystem. " +
+          "Skipping on block location reordering");
       return false;
     }
 
@@ -243,7 +241,8 @@ public class HFileSystem extends FilterF
 
       ClientProtocol cp1 = createReorderingProxy(namenode, lrb, conf);
       nf.set(dfsc, cp1);
-      LOG.info("Added intercepting call to namenode#getBlockLocations");
+      LOG.info("Added intercepting call to namenode#getBlockLocations so can do block reordering" +
+        " using class " + lrb.getClass());
     } catch (NoSuchFieldException e) {
       LOG.warn("Can't modify the DFSClient#namenode field to add the location reorder.", e);
       return false;

Modified: hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java (original)
+++ hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/CacheConfig.java Fri May  3 03:58:33 2013
@@ -347,8 +347,7 @@ public class CacheConfig {
    * @param conf  The current configuration.
    * @return The block cache or <code>null</code>.
    */
-  private static synchronized BlockCache instantiateBlockCache(
-      Configuration conf) {
+  private static synchronized BlockCache instantiateBlockCache(Configuration conf) {
     if (globalBlockCache != null) return globalBlockCache;
     if (blockCacheDisabled) return null;
 
@@ -366,14 +365,12 @@ public class CacheConfig {
     // Calculate the amount of heap to give the heap.
     MemoryUsage mu = ManagementFactory.getMemoryMXBean().getHeapMemoryUsage();
     long lruCacheSize = (long) (mu.getMax() * cachePercentage);
-    int blockSize = conf.getInt("hbase.offheapcache.minblocksize",
-        HConstants.DEFAULT_BLOCKSIZE);
+    int blockSize = conf.getInt("hbase.offheapcache.minblocksize", HConstants.DEFAULT_BLOCKSIZE);
     long offHeapCacheSize =
       (long) (conf.getFloat("hbase.offheapcache.percentage", (float) 0) *
           DirectMemoryUtils.getDirectMemorySize());
     if (offHeapCacheSize <= 0) {
-      String bucketCacheIOEngineName = conf
-          .get(BUCKET_CACHE_IOENGINE_KEY, null);
+      String bucketCacheIOEngineName = conf.get(BUCKET_CACHE_IOENGINE_KEY, null);
       float bucketCachePercentage = conf.getFloat(BUCKET_CACHE_SIZE_KEY, 0F);
       // A percentage of max heap size or a absolute value with unit megabytes
       long bucketCacheSize = (long) (bucketCachePercentage < 1 ? mu.getMax()
@@ -407,10 +404,9 @@ public class CacheConfig {
           throw new RuntimeException(ioex);
         }
       }
-      LOG.info("Allocating LruBlockCache with maximum size "
-          + StringUtils.humanReadableInt(lruCacheSize));
-      LruBlockCache lruCache = new LruBlockCache(lruCacheSize,
-          StoreFile.DEFAULT_BLOCKSIZE_SMALL);
+      LOG.info("Allocating LruBlockCache with maximum size " +
+        StringUtils.humanReadableInt(lruCacheSize));
+      LruBlockCache lruCache = new LruBlockCache(lruCacheSize, StoreFile.DEFAULT_BLOCKSIZE_SMALL);
       lruCache.setVictimCache(bucketCache);
       if (bucketCache != null && combinedWithLru) {
         globalBlockCache = new CombinedBlockCache(lruCache, bucketCache);

Added: hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/EmptyServiceNameException.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/EmptyServiceNameException.java?rev=1478639&view=auto
==============================================================================
--- hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/EmptyServiceNameException.java (added)
+++ hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/EmptyServiceNameException.java Fri May  3 03:58:33 2013
@@ -0,0 +1,21 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hbase.ipc;
+
+@SuppressWarnings("serial")
+public class EmptyServiceNameException extends FatalConnectionException {}
\ No newline at end of file

Modified: hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperImpl.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperImpl.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperImpl.java (original)
+++ hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/MetricsHBaseServerWrapperImpl.java Fri May  3 03:58:33 2013
@@ -21,9 +21,9 @@ package org.apache.hadoop.hbase.ipc;
 
 public class MetricsHBaseServerWrapperImpl implements MetricsHBaseServerWrapper {
 
-  private HBaseServer server;
+  private RpcServer server;
 
-  MetricsHBaseServerWrapperImpl(HBaseServer server) {
+  MetricsHBaseServerWrapperImpl(RpcServer server) {
     this.server = server;
   }
 

Modified: hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RequestContext.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RequestContext.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RequestContext.java (original)
+++ hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RequestContext.java Fri May  3 03:58:33 2013
@@ -21,9 +21,10 @@
 package org.apache.hadoop.hbase.ipc;
 
 import org.apache.hadoop.classification.InterfaceAudience;
-import org.apache.hadoop.hbase.IpcProtocol;
 import org.apache.hadoop.hbase.security.User;
 
+import com.google.protobuf.BlockingService;
+
 import java.net.InetAddress;
 
 /**
@@ -87,15 +88,14 @@ public class RequestContext {
    * Initializes the client credentials for the current request.
    * @param user
    * @param remoteAddress
-   * @param protocol
+   * @param service
    */
   public static void set(User user,
-      InetAddress remoteAddress,
-      Class<? extends IpcProtocol> protocol) {
+      InetAddress remoteAddress, BlockingService service) {
     RequestContext ctx = instance.get();
     ctx.user = user;
     ctx.remoteAddress = remoteAddress;
-    ctx.protocol = protocol;
+    ctx.service = service;
     ctx.inRequest = true;
   }
 
@@ -106,21 +106,20 @@ public class RequestContext {
     RequestContext ctx = instance.get();
     ctx.user = null;
     ctx.remoteAddress = null;
-    ctx.protocol = null;
+    ctx.service = null;
     ctx.inRequest = false;
   }
 
   private User user;
   private InetAddress remoteAddress;
-  private Class<? extends IpcProtocol> protocol;
+  private BlockingService service;
   // indicates we're within a RPC request invocation
   private boolean inRequest;
 
-  private RequestContext(User user, InetAddress remoteAddr,
-      Class<? extends IpcProtocol> protocol) {
+  private RequestContext(User user, InetAddress remoteAddr, BlockingService service) {
     this.user = user;
     this.remoteAddress = remoteAddr;
-    this.protocol = protocol;
+    this.service = service;
   }
 
   public User getUser() {
@@ -131,8 +130,8 @@ public class RequestContext {
     return remoteAddress;
   }
 
-  public Class<? extends IpcProtocol> getProtocol() {
-    return protocol;
+  public BlockingService getService() {
+    return this.service;
   }
 
   public boolean isInRequest() {

Modified: hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java (original)
+++ hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/ResponseFlag.java Fri May  3 03:58:33 2013
@@ -21,7 +21,7 @@ import org.apache.hadoop.classification.
 
 /**
  * Utility for managing the flag byte passed in response to a
- * {@link HBaseServer.Call}
+ * {@link RpcServer.Call}
  */
 @InterfaceAudience.Private
 class ResponseFlag {

Modified: hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java
URL: http://svn.apache.org/viewvc/hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java?rev=1478639&r1=1478638&r2=1478639&view=diff
==============================================================================
--- hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java (original)
+++ hbase/branches/0.95/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcCallContext.java Fri May  3 03:58:33 2013
@@ -20,12 +20,10 @@ package org.apache.hadoop.hbase.ipc;
 import org.apache.hadoop.hbase.exceptions.CallerDisconnectedException;
 
 public interface RpcCallContext extends Delayable {
-
   /**
    * Throw an exception if the caller who made this IPC call has disconnected.
    * If called from outside the context of IPC, this does nothing.
    * @throws CallerDisconnectedException
    */
   void throwExceptionIfCallerDisconnected() throws CallerDisconnectedException;
-
 }