You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hbase.apache.org by st...@apache.org on 2018/02/03 06:42:39 UTC

hbase git commit: HBASE-19919 Tidying up logging

Repository: hbase
Updated Branches:
  refs/heads/branch-2 897afb7d1 -> bac468734


HBASE-19919 Tidying up logging


Project: http://git-wip-us.apache.org/repos/asf/hbase/repo
Commit: http://git-wip-us.apache.org/repos/asf/hbase/commit/bac46873
Tree: http://git-wip-us.apache.org/repos/asf/hbase/tree/bac46873
Diff: http://git-wip-us.apache.org/repos/asf/hbase/diff/bac46873

Branch: refs/heads/branch-2
Commit: bac4687345bdc975501a898b4cf0930784f29c66
Parents: 897afb7
Author: Michael Stack <st...@apache.org>
Authored: Fri Feb 2 00:23:37 2018 -0800
Committer: Michael Stack <st...@apache.org>
Committed: Fri Feb 2 22:42:30 2018 -0800

----------------------------------------------------------------------
 .../hadoop/hbase/ipc/NettyRpcConnection.java    |  4 +-
 .../hbase/zookeeper/ReadOnlyZKClient.java       |  8 +--
 .../org/apache/hadoop/hbase/ChoreService.java   |  2 +-
 .../org/apache/hadoop/hbase/net/Address.java    | 19 +++++++
 .../apache/hadoop/hbase/net/TestAddress.java    | 50 +++++++++++++++++
 .../hbase/procedure2/ProcedureExecutor.java     | 16 ++----
 .../procedure2/RemoteProcedureDispatcher.java   |  2 +-
 .../hbase/procedure2/StateMachineProcedure.java |  4 +-
 .../procedure2/store/wal/WALProcedureStore.java |  8 +--
 .../replication/ReplicationPeersZKImpl.java     |  2 +-
 .../ZKSplitLogManagerCoordination.java          | 58 ++++++++++----------
 .../hbase/coprocessor/CoprocessorHost.java      |  3 +-
 .../apache/hadoop/hbase/ipc/RpcExecutor.java    |  4 +-
 .../hbase/master/ActiveMasterManager.java       |  4 +-
 .../org/apache/hadoop/hbase/master/HMaster.java | 11 ++--
 .../hbase/master/MasterCoprocessorHost.java     |  2 +-
 .../hadoop/hbase/master/ServerManager.java      |  8 +--
 .../hadoop/hbase/master/SplitLogManager.java    |  4 +-
 .../hadoop/hbase/master/TableStateManager.java  |  5 +-
 .../master/assignment/AssignmentManager.java    |  8 +--
 .../hbase/master/cleaner/CleanerChore.java      | 19 ++-----
 .../hbase/master/cleaner/HFileCleaner.java      |  4 +-
 .../hadoop/hbase/master/cleaner/LogCleaner.java |  2 +-
 .../master/procedure/RSProcedureDispatcher.java |  5 +-
 .../hbase/procedure/ZKProcedureCoordinator.java |  2 +-
 .../hadoop/hbase/procedure/ZKProcedureUtil.java |  3 +-
 .../hadoop/hbase/regionserver/ChunkCreator.java | 14 ++---
 .../hbase/regionserver/HRegionServer.java       |  6 +-
 .../hbase/regionserver/RSRpcServices.java       |  7 ++-
 .../master/ReplicationLogCleaner.java           |  4 +-
 .../master/ReplicationPeerConfigUpgrader.java   |  2 +-
 .../hbase/security/access/AccessController.java |  2 +-
 .../security/access/ZKPermissionWatcher.java    |  2 +-
 .../hadoop/hbase/util/FSTableDescriptors.java   | 12 ++--
 .../hbase/zookeeper/RecoverableZooKeeper.java   |  2 +-
 35 files changed, 181 insertions(+), 127 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcConnection.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcConnection.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcConnection.java
index d745716..f8a4557 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcConnection.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/NettyRpcConnection.java
@@ -251,9 +251,7 @@ class NettyRpcConnection extends RpcConnection {
   }
 
   private void connect() {
-    if (LOG.isDebugEnabled()) {
-      LOG.debug("Connecting to " + remoteId.address);
-    }
+    LOG.debug("Connecting to {}", remoteId.address);
 
     this.channel = new Bootstrap().group(rpcClient.group).channel(rpcClient.channelClass)
         .option(ChannelOption.TCP_NODELAY, rpcClient.isTcpNoDelay())

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.java
----------------------------------------------------------------------
diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.java
index 275fafb..e41f04a 100644
--- a/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.java
+++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/zookeeper/ReadOnlyZKClient.java
@@ -131,8 +131,8 @@ public final class ReadOnlyZKClient implements Closeable {
         conf.getInt(RECOVERY_RETRY_INTERVAL_MILLIS, DEFAULT_RECOVERY_RETRY_INTERVAL_MILLIS);
     this.keepAliveTimeMs = conf.getInt(KEEPALIVE_MILLIS, DEFAULT_KEEPALIVE_MILLIS);
     LOG.info(
-      "Start read only zookeeper connection {} to {}, " + "session timeout {} ms, retries {}, " +
-        "retry interval {} ms, keep alive {} ms",
+      "Connect {} to {} with session timeout={}ms, retries {}, " +
+        "retry interval {}ms, keepAlive={}ms",
       getId(), connectString, sessionTimeoutMs, maxRetries, retryIntervalMs, keepAliveTimeMs);
     Threads.setDaemonThreadRunning(new Thread(this::run),
       "ReadOnlyZKClient-" + connectString + "@" + getId());
@@ -313,9 +313,7 @@ public final class ReadOnlyZKClient implements Closeable {
       }
       if (task == null) {
         if (pendingRequests == 0) {
-          LOG.debug(
-            "{} to {} no activities for {} ms, close active connection. " +
-              "Will reconnect next time when there are new requests",
+          LOG.debug("{} to {} inactive for {}ms; closing (Will reconnect when new requests)",
             getId(), connectString, keepAliveTimeMs);
           closeZk();
         }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
index 3a1a8e9..85d6131 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/ChoreService.java
@@ -247,7 +247,7 @@ public class ChoreService implements ChoreServicer {
    */
   static class ChoreServiceThreadFactory implements ThreadFactory {
     private final String threadPrefix;
-    private final static String THREAD_NAME_SUFFIX = "_Chore_";
+    private final static String THREAD_NAME_SUFFIX = ".Chore.";
     private AtomicInteger threadNumber = new AtomicInteger(1);
 
     /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-common/src/main/java/org/apache/hadoop/hbase/net/Address.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/net/Address.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/net/Address.java
index 1b12a66..9d7f65c 100644
--- a/hbase-common/src/main/java/org/apache/hadoop/hbase/net/Address.java
+++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/net/Address.java
@@ -17,6 +17,7 @@
  */
 package org.apache.hadoop.hbase.net;
 
+import org.apache.commons.lang.StringUtils;
 import org.apache.yetus.audience.InterfaceAudience;
 
 import org.apache.hbase.thirdparty.com.google.common.net.HostAndPort;
@@ -58,6 +59,24 @@ public class Address implements Comparable<Address> {
     return this.hostAndPort.toString();
   }
 
+  /**
+   * If hostname is a.b.c and the port is 123, return a:123 instead of a.b.c:123.
+   * @return if host looks like it is resolved -- not an IP -- then strip the domain portion
+   * otherwise returns same as {@link #toString()}}
+   */
+  public String toStringWithoutDomain() {
+    String hostname = getHostname();
+    String [] parts = hostname.split("\\.");
+    if (parts.length > 1) {
+      for (String part: parts) {
+        if (!StringUtils.isNumeric(part)) {
+          return Address.fromParts(parts[0], getPort()).toString();
+        }
+      }
+    }
+    return toString();
+  }
+
   @Override
   // Don't use HostAndPort equals... It is wonky including
   // ipv6 brackets

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-common/src/test/java/org/apache/hadoop/hbase/net/TestAddress.java
----------------------------------------------------------------------
diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/net/TestAddress.java b/hbase-common/src/test/java/org/apache/hadoop/hbase/net/TestAddress.java
new file mode 100644
index 0000000..22c7940
--- /dev/null
+++ b/hbase-common/src/test/java/org/apache/hadoop/hbase/net/TestAddress.java
@@ -0,0 +1,50 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *     http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.hadoop.hbase.net;
+
+
+import org.apache.hadoop.hbase.HBaseClassTestRule;
+import org.apache.hadoop.hbase.testclassification.MiscTests;
+import org.apache.hadoop.hbase.testclassification.SmallTests;
+import org.junit.Assert;
+import org.junit.ClassRule;
+import org.junit.Test;
+import org.junit.experimental.categories.Category;
+
+import static org.junit.Assert.assertEquals;
+
+@Category({ MiscTests.class, SmallTests.class })
+public class TestAddress {
+  @ClassRule
+  public static final HBaseClassTestRule CLASS_RULE =
+      HBaseClassTestRule.forClass(TestAddress.class);
+
+  @Test
+  public void testGetHostWithoutDomain() {
+    assertEquals("a:123",
+        Address.fromParts("a.b.c", 123).toStringWithoutDomain());
+    assertEquals("1:123",
+        Address.fromParts("1.b.c", 123).toStringWithoutDomain());
+    assertEquals("123.456.789.1:123",
+        Address.fromParts("123.456.789.1", 123).toStringWithoutDomain());
+    assertEquals("[2001:db8::1]:80",
+        Address.fromParts("[2001:db8::1]:", 80).toStringWithoutDomain());
+    assertEquals("[2001:db8::1]:80",
+        Address.fromParts("[2001:db8::1]:", 80).toStringWithoutDomain());
+  }
+}

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
index 4a61299..b1ff426 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/ProcedureExecutor.java
@@ -230,9 +230,7 @@ public class ProcedureExecutor<TEnvironment> {
             nonceKeysToProcIdsMap.remove(nonceKey);
           }
           it.remove();
-          if (debugEnabled) {
-            LOG.debug("Evict completed " + proc);
-          }
+          LOG.trace("Evict completed {}", proc);
         }
       }
       if (batchCount > 0) {
@@ -316,7 +314,6 @@ public class ProcedureExecutor<TEnvironment> {
       @Override
       public void setMaxProcId(long maxProcId) {
         assert lastProcId.get() < 0 : "expected only one call to setMaxProcId()";
-        LOG.debug("Load max pid=" + maxProcId);
         lastProcId.set(maxProcId);
       }
 
@@ -512,10 +509,10 @@ public class ProcedureExecutor<TEnvironment> {
     // We have numThreads executor + one timer thread used for timing out
     // procedures and triggering periodic procedures.
     this.corePoolSize = numThreads;
-    LOG.info("Starting ProcedureExecutor Worker threads (ProcedureExecutorWorker)=" + corePoolSize);
+    LOG.info("Starting {} Workers (bigger of cpus/4 or 16)", corePoolSize);
 
     // Create the Thread Group for the executors
-    threadGroup = new ThreadGroup("ProcedureExecutorWorkerGroup");
+    threadGroup = new ThreadGroup("PEWorkerGroup");
 
     // Create the timeout executor
     timeoutExecutor = new TimeoutExecutorThread(threadGroup);
@@ -533,7 +530,7 @@ public class ProcedureExecutor<TEnvironment> {
     st = EnvironmentEdgeManager.currentTime();
     store.recoverLease();
     et = EnvironmentEdgeManager.currentTime();
-    LOG.info(String.format("Recover store (%s) lease: %s",
+    LOG.info(String.format("Recovered %s lease in %s",
       store.getClass().getSimpleName(), StringUtils.humanTimeDiff(et - st)));
 
     // start the procedure scheduler
@@ -547,7 +544,7 @@ public class ProcedureExecutor<TEnvironment> {
     st = EnvironmentEdgeManager.currentTime();
     load(abortOnCorruption);
     et = EnvironmentEdgeManager.currentTime();
-    LOG.info(String.format("Load store (%s): %s",
+    LOG.info(String.format("Loaded %s in %s, um pid=",
       store.getClass().getSimpleName(), StringUtils.humanTimeDiff(et - st)));
 
     // Start the executors. Here we must have the lastProcId set.
@@ -1612,7 +1609,6 @@ public class ProcedureExecutor<TEnvironment> {
     }
 
     // If this procedure is the last child awake the parent procedure
-    LOG.info("Finish suprocedure " + procedure);
     if (parent.tryRunnable()) {
       // If we succeeded in making the parent runnable -- i.e. all of its
       // children have completed, move parent to front of the queue.
@@ -1711,7 +1707,7 @@ public class ProcedureExecutor<TEnvironment> {
     private Procedure activeProcedure;
 
     public WorkerThread(final ThreadGroup group) {
-      super(group, "ProcedureExecutorWorker-" + workerId.incrementAndGet());
+      super(group, "PEWorker-" + workerId.incrementAndGet());
       setDaemon(true);
     }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java
index 72a2ebb..861e3b2 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/RemoteProcedureDispatcher.java
@@ -94,7 +94,7 @@ public abstract class RemoteProcedureDispatcher<TEnv, TRemote extends Comparable
       return false;
     }
 
-    LOG.info("Starting procedure remote dispatcher; threads=" + this.corePoolSize +
+    LOG.info("Started, threads=" + this.corePoolSize +
       ", queueMaxSize=" + this.queueMaxSize + ", operationDelay=" + this.operationDelay);
 
     // Create the timeout executor

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java
index ade07cc..20bba58 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/StateMachineProcedure.java
@@ -176,9 +176,7 @@ public abstract class StateMachineProcedure<TEnvironment, TState>
         this.cycles++;
       }
 
-      if (LOG.isDebugEnabled()) {
-        LOG.debug(toString());
-      }
+      LOG.trace("{}", toString());
       stateFlow = executeFromState(env, state);
       if (!hasMoreState()) setNextState(EOF_STATE);
       if (subProcList != null && !subProcList.isEmpty()) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java
----------------------------------------------------------------------
diff --git a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java
index 5f9e7f6..e21f64e 100644
--- a/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java
+++ b/hbase-procedure/src/main/java/org/apache/hadoop/hbase/procedure2/store/wal/WALProcedureStore.java
@@ -358,7 +358,7 @@ public class WALProcedureStore extends ProcedureStoreBase {
   public void recoverLease() throws IOException {
     lock.lock();
     try {
-      LOG.info("Starting WAL Procedure Store lease recovery");
+      LOG.trace("Starting WAL Procedure Store lease recovery");
       FileStatus[] oldLogs = getLogFiles();
       while (isRunning()) {
         // Get Log-MaxID and recover lease on old logs
@@ -387,7 +387,7 @@ public class WALProcedureStore extends ProcedureStoreBase {
           continue;
         }
 
-        LOG.info("Lease acquired for flushLogId: " + flushLogId);
+        LOG.trace("Lease acquired for flushLogId={}", flushLogId);
         break;
       }
     } finally {
@@ -405,9 +405,7 @@ public class WALProcedureStore extends ProcedureStoreBase {
 
       // Nothing to do, If we have only the current log.
       if (logs.size() == 1) {
-        if (LOG.isDebugEnabled()) {
-          LOG.debug("No state logs to replay.");
-        }
+        LOG.trace("No state logs to replay.");
         loader.setMaxProcId(0);
         return;
       }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-replication/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeersZKImpl.java
----------------------------------------------------------------------
diff --git a/hbase-replication/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeersZKImpl.java b/hbase-replication/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeersZKImpl.java
index 4e469ed..8e9d98e 100644
--- a/hbase-replication/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeersZKImpl.java
+++ b/hbase-replication/src/main/java/org/apache/hadoop/hbase/replication/ReplicationPeersZKImpl.java
@@ -451,7 +451,7 @@ public class ReplicationPeersZKImpl extends ReplicationStateZKBase implements Re
     ReplicationPeerZKImpl previous =
       ((ConcurrentMap<String, ReplicationPeerZKImpl>) peerClusters).putIfAbsent(peerId, peer);
     if (previous == null) {
-      LOG.info("Added new peer cluster=" + peer.getPeerConfig().getClusterKey());
+      LOG.info("Added peer cluster=" + peer.getPeerConfig().getClusterKey());
     } else {
       LOG.info("Peer already present, " + previous.getPeerConfig().getClusterKey() +
         ", new cluster=" + peer.getPeerConfig().getClusterKey());

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZKSplitLogManagerCoordination.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZKSplitLogManagerCoordination.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZKSplitLogManagerCoordination.java
index 4e0c973..d2777a0 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZKSplitLogManagerCoordination.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coordination/ZKSplitLogManagerCoordination.java
@@ -151,7 +151,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
     }
     Task task = findOrCreateOrphanTask(path);
     if (task.isOrphan() && (task.incarnation.get() == 0)) {
-      LOG.info("resubmitting unassigned orphan task " + path);
+      LOG.info("Resubmitting unassigned orphan task " + path);
       // ignore failure to resubmit. The timeout-monitor will handle it later
       // albeit in a more crude fashion
       resubmitTask(path, task, FORCE);
@@ -202,7 +202,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
       SplitLogCounters.tot_mgr_resubmit_force.increment();
       version = -1;
     }
-    LOG.info("resubmitting task " + path);
+    LOG.info("Resubmitting task " + path);
     task.incarnation.incrementAndGet();
     boolean result = resubmit(path, version);
     if (!result) {
@@ -280,7 +280,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
         SplitLogCounters.tot_mgr_rescan_deleted.increment();
       }
       SplitLogCounters.tot_mgr_missing_state_in_delete.increment();
-      LOG.debug("deleted task without in memory state " + path);
+      LOG.debug("Deleted task without in memory state " + path);
       return;
     }
     synchronized (task) {
@@ -328,13 +328,13 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
   }
 
   private void createNodeSuccess(String path) {
-    LOG.debug("put up splitlog task at znode " + path);
+    LOG.debug("Put up splitlog task at znode " + path);
     getDataSetWatch(path, zkretries);
   }
 
   private void createNodeFailure(String path) {
     // TODO the Manager should split the log locally instead of giving up
-    LOG.warn("failed to create task node" + path);
+    LOG.warn("Failed to create task node " + path);
     setDone(path, FAILURE);
   }
 
@@ -360,15 +360,15 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
     data = ZKMetadata.removeMetaData(data);
     SplitLogTask slt = SplitLogTask.parseFrom(data);
     if (slt.isUnassigned()) {
-      LOG.debug("task not yet acquired " + path + " ver = " + version);
+      LOG.debug("Task not yet acquired " + path + ", ver=" + version);
       handleUnassignedTask(path);
     } else if (slt.isOwned()) {
       heartbeat(path, version, slt.getServerName());
     } else if (slt.isResigned()) {
-      LOG.info("task " + path + " entered state: " + slt.toString());
+      LOG.info("Task " + path + " entered state=" + slt.toString());
       resubmitOrFail(path, FORCE);
     } else if (slt.isDone()) {
-      LOG.info("task " + path + " entered state: " + slt.toString());
+      LOG.info("Task " + path + " entered state=" + slt.toString());
       if (taskFinisher != null && !ZKSplitLog.isRescanNode(watcher, path)) {
         if (taskFinisher.finish(slt.getServerName(), ZKSplitLog.getFileName(path)) == Status.DONE) {
           setDone(path, SUCCESS);
@@ -379,7 +379,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
         setDone(path, SUCCESS);
       }
     } else if (slt.isErr()) {
-      LOG.info("task " + path + " entered state: " + slt.toString());
+      LOG.info("Task " + path + " entered state=" + slt.toString());
       resubmitOrFail(path, CHECK);
     } else {
       LOG.error(HBaseMarkers.FATAL, "logic error - unexpected zk state for path = "
@@ -395,7 +395,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
   }
 
   private void getDataSetWatchFailure(String path) {
-    LOG.warn("failed to set data watch " + path);
+    LOG.warn("Failed to set data watch " + path);
     setDone(path, FAILURE);
   }
 
@@ -404,7 +404,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
     if (task == null) {
       if (!ZKSplitLog.isRescanNode(watcher, path)) {
         SplitLogCounters.tot_mgr_unacquired_orphan_done.increment();
-        LOG.debug("unacquired orphan task is done " + path);
+        LOG.debug("Unacquired orphan task is done " + path);
       }
     } else {
       synchronized (task) {
@@ -441,7 +441,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
 
   private Task findOrCreateOrphanTask(String path) {
     return computeIfAbsent(details.getTasks(), path, Task::new, () -> {
-      LOG.info("creating orphan task " + path);
+      LOG.info("Creating orphan task " + path);
       SplitLogCounters.tot_mgr_orphan_task_acquired.increment();
     });
   }
@@ -450,7 +450,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
     Task task = findOrCreateOrphanTask(path);
     if (new_version != task.last_version) {
       if (task.isUnassigned()) {
-        LOG.info("task " + path + " acquired by " + workerName);
+        LOG.info("Task " + path + " acquired by " + workerName);
       }
       task.heartbeat(EnvironmentEdgeManager.currentTime(), new_version, workerName);
       SplitLogCounters.tot_mgr_heartbeat.increment();
@@ -468,11 +468,11 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
     try {
       orphans = ZKUtil.listChildrenNoWatch(this.watcher, this.watcher.znodePaths.splitLogZNode);
       if (orphans == null) {
-        LOG.warn("could not get children of " + this.watcher.znodePaths.splitLogZNode);
+        LOG.warn("Could not get children of " + this.watcher.znodePaths.splitLogZNode);
         return;
       }
     } catch (KeeperException e) {
-      LOG.warn("could not get children of " + this.watcher.znodePaths.splitLogZNode + " "
+      LOG.warn("Could not get children of " + this.watcher.znodePaths.splitLogZNode + " "
           + StringUtils.stringifyException(e));
       return;
     }
@@ -483,9 +483,9 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
       String nodepath = ZNodePaths.joinZNode(watcher.znodePaths.splitLogZNode, path);
       if (ZKSplitLog.isRescanNode(watcher, nodepath)) {
         rescan_nodes++;
-        LOG.debug("found orphan rescan node " + path);
+        LOG.debug("Found orphan rescan node " + path);
       } else {
-        LOG.info("found orphan task " + path);
+        LOG.info("Found orphan task " + path);
       }
       getDataSetWatch(nodepath, zkretries);
     }
@@ -511,11 +511,11 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
       SplitLogTask slt =
           new SplitLogTask.Unassigned(this.details.getServerName());
       if (ZKUtil.setData(this.watcher, path, slt.toByteArray(), version) == false) {
-        LOG.debug("failed to resubmit task " + path + " version changed");
+        LOG.debug("Failed to resubmit task " + path + " version changed");
         return false;
       }
     } catch (NoNodeException e) {
-      LOG.warn("failed to resubmit because znode doesn't exist " + path
+      LOG.warn("Failed to resubmit because znode doesn't exist " + path
           + " task done (or forced done by removing the znode)");
       try {
         getDataSetWatchSuccess(path, null, Integer.MIN_VALUE);
@@ -525,11 +525,11 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
       }
       return false;
     } catch (KeeperException.BadVersionException e) {
-      LOG.debug("failed to resubmit task " + path + " version changed");
+      LOG.debug("Failed to resubmit task " + path + " version changed");
       return false;
     } catch (KeeperException e) {
       SplitLogCounters.tot_mgr_resubmit_failed.increment();
-      LOG.warn("failed to resubmit " + path, e);
+      LOG.warn("Failed to resubmit " + path, e);
       return false;
     }
     return true;
@@ -590,11 +590,11 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
           // this manager are get-znode-data, task-finisher and delete-znode.
           // And all code pieces correctly handle the case of suddenly
           // disappearing task-znode.
-          LOG.debug("found pre-existing znode " + path);
+          LOG.debug("Found pre-existing znode " + path);
           SplitLogCounters.tot_mgr_node_already_exists.increment();
         } else {
           Long retry_count = (Long) ctx;
-          LOG.warn("create rc =" + KeeperException.Code.get(rc) + " for " + path
+          LOG.warn("Create rc=" + KeeperException.Code.get(rc) + " for " + path
               + " remaining retries=" + retry_count);
           if (retry_count == 0) {
             SplitLogCounters.tot_mgr_node_create_err.increment();
@@ -625,7 +625,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
         }
         if (rc == KeeperException.Code.NONODE.intValue()) {
           SplitLogCounters.tot_mgr_get_data_nonode.increment();
-          LOG.warn("task znode " + path + " vanished or not created yet.");
+          LOG.warn("Task znode " + path + " vanished or not created yet.");
           // ignore since we should not end up in a case where there is in-memory task,
           // but no znode. The only case is between the time task is created in-memory
           // and the znode is created. See HBASE-11217.
@@ -634,11 +634,11 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
         Long retry_count = (Long) ctx;
 
         if (retry_count < 0) {
-          LOG.warn("getdata rc = " + KeeperException.Code.get(rc) + " " + path
+          LOG.warn("Getdata rc=" + KeeperException.Code.get(rc) + " " + path
               + ". Ignoring error. No error handling. No retrying.");
           return;
         }
-        LOG.warn("getdata rc = " + KeeperException.Code.get(rc) + " " + path
+        LOG.warn("Getdata rc=" + KeeperException.Code.get(rc) + " " + path
             + " remaining retries=" + retry_count);
         if (retry_count == 0) {
           SplitLogCounters.tot_mgr_get_data_err.increment();
@@ -675,10 +675,10 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
         if (rc != KeeperException.Code.NONODE.intValue()) {
           SplitLogCounters.tot_mgr_node_delete_err.increment();
           Long retry_count = (Long) ctx;
-          LOG.warn("delete rc=" + KeeperException.Code.get(rc) + " for " + path
+          LOG.warn("Delete rc=" + KeeperException.Code.get(rc) + " for " + path
               + " remaining retries=" + retry_count);
           if (retry_count == 0) {
-            LOG.warn("delete failed " + path);
+            LOG.warn("Delete failed " + path);
             details.getFailedDeletions().add(path);
             deleteNodeFailure(path);
           } else {
@@ -691,7 +691,7 @@ public class ZKSplitLogManagerCoordination extends ZKListener implements
               + " in earlier retry rounds. zkretries = " + ctx);
         }
       } else {
-        LOG.debug("deleted " + path);
+        LOG.debug("Deleted " + path);
       }
       deleteNodeSuccess(path);
     }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java
index 05ac9f6..4c56056 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/CoprocessorHost.java
@@ -157,8 +157,7 @@ public abstract class CoprocessorHost<C extends Coprocessor, E extends Coprocess
         E env = checkAndLoadInstance(implClass, priority, conf);
         if (env != null) {
           this.coprocEnvironments.add(env);
-          LOG.info(
-              "System coprocessor " + className + " was loaded " + "successfully with priority (" + priority + ").");
+          LOG.info("System coprocessor {} loaded, priority={}.", className, priority);
           ++priority;
         }
       } catch (Throwable t) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcExecutor.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcExecutor.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcExecutor.java
index fa30d0d..5572725 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcExecutor.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/ipc/RpcExecutor.java
@@ -251,8 +251,6 @@ public abstract class RpcExecutor {
     double handlerFailureThreshhold = conf == null ? 1.0 : conf.getDouble(
       HConstants.REGION_SERVER_HANDLER_ABORT_ON_ERROR_PERCENT,
       HConstants.DEFAULT_REGION_SERVER_HANDLER_ABORT_ON_ERROR_PERCENT);
-    LOG.debug("Started " + handlers.size() + " " + threadPrefix +
-        " handlers, qsize=" + qsize + " on port=" + port);
     for (int i = 0; i < numHandlers; i++) {
       final int index = qindex + (i % qsize);
       String name = "RpcServer." + threadPrefix + ".handler=" + handlers.size() + ",queue=" + index
@@ -262,6 +260,8 @@ public abstract class RpcExecutor {
       handler.start();
       handlers.add(handler);
     }
+    LOG.debug("Started " + handlers.size() + " " + threadPrefix +
+        " handlers, queues=" + qsize + ", port=" + port);
   }
 
   /**

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
index 1cc519b..78cb024 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ActiveMasterManager.java
@@ -125,7 +125,7 @@ public class ActiveMasterManager extends ZKListener {
       synchronized(clusterHasActiveMaster) {
         if (ZKUtil.watchAndCheckExists(watcher, watcher.znodePaths.masterAddressZNode)) {
           // A master node exists, there is an active master
-          LOG.debug("A master is now available");
+          LOG.trace("A master is now available");
           clusterHasActiveMaster.set(true);
         } else {
           // Node is no longer there, cluster does not have an active master
@@ -178,7 +178,7 @@ public class ActiveMasterManager extends ZKListener {
           // We are the master, return
           startupStatus.setStatus("Successfully registered as active master.");
           this.clusterHasActiveMaster.set(true);
-          LOG.info("Registered Active Master=" + this.sn);
+          LOG.info("Registered as active master=" + this.sn);
           return true;
         }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
index a75c9fe..378b338 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
@@ -745,7 +745,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     boolean wasUp = this.clusterStatusTracker.isClusterUp();
     if (!wasUp) this.clusterStatusTracker.setClusterUp();
 
-    LOG.info("Server active/primary master=" + this.serverName +
+    LOG.info("Active/primary master=" + this.serverName +
         ", sessionid=0x" +
         Long.toHexString(this.zooKeeper.getRecoverableZooKeeper().getSessionId()) +
         ", setting cluster-up flag (Was=" + wasUp + ")");
@@ -1145,7 +1145,7 @@ public class HMaster extends HRegionServer implements MasterServices {
    startProcedureExecutor();
 
    // Start log cleaner thread
-   int cleanerInterval = conf.getInt("hbase.master.cleaner.interval", 60 * 1000);
+   int cleanerInterval = conf.getInt("hbase.master.cleaner.interval", 600 * 1000);
    this.logCleaner =
       new LogCleaner(cleanerInterval,
          this, conf, getMasterWalManager().getFileSystem(),
@@ -1222,9 +1222,10 @@ public class HMaster extends HRegionServer implements MasterServices {
     procedureExecutor = new ProcedureExecutor<>(conf, procEnv, procedureStore, procedureScheduler);
     configurationManager.registerObserver(procEnv);
 
+    int cpus = Runtime.getRuntime().availableProcessors();
     final int numThreads = conf.getInt(MasterProcedureConstants.MASTER_PROCEDURE_THREADS,
-        Math.max(Runtime.getRuntime().availableProcessors(),
-          MasterProcedureConstants.DEFAULT_MIN_MASTER_PROCEDURE_THREADS));
+        Math.max((cpus > 0? cpus/4: 0),
+            MasterProcedureConstants.DEFAULT_MIN_MASTER_PROCEDURE_THREADS));
     final boolean abortOnCorruption = conf.getBoolean(
         MasterProcedureConstants.EXECUTOR_ABORT_ON_CORRUPTION,
         MasterProcedureConstants.DEFAULT_EXECUTOR_ABORT_ON_CORRUPTION);
@@ -3549,7 +3550,7 @@ public class HMaster extends HRegionServer implements MasterServices {
     LOG.info("Running RecoverMetaProcedure to ensure proper hbase:meta deploy.");
     long procId = procedureExecutor.submitProcedure(new RecoverMetaProcedure(null, true, latch));
     latch.await();
-    LOG.info("hbase:meta (default replica) deployed at=" +
+    LOG.info("hbase:meta deployed at=" +
         getMetaTableLocator().getMetaRegionLocation(getZooKeeper()));
     return assignmentManager.isMetaInitialized();
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
index 8396145..7f17227 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/MasterCoprocessorHost.java
@@ -150,7 +150,7 @@ public class MasterCoprocessorHost
     // only one MasterCoprocessorHost instance in the master process
     boolean coprocessorsEnabled = conf.getBoolean(COPROCESSORS_ENABLED_CONF_KEY,
       DEFAULT_COPROCESSORS_ENABLED);
-    LOG.info("System coprocessor loading is " + (coprocessorsEnabled ? "enabled" : "disabled"));
+    LOG.trace("System coprocessor loading is {}",  (coprocessorsEnabled ? "enabled" : "disabled"));
     loadSystemCoprocessors(conf, MASTER_COPROCESSOR_CONF_KEY);
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
index aa75091..dbcce1d 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/ServerManager.java
@@ -424,7 +424,7 @@ public class ServerManager {
    */
   @VisibleForTesting
   void recordNewServerWithLock(final ServerName serverName, final ServerLoad sl) {
-    LOG.info("Registering server=" + serverName);
+    LOG.info("Registering regionserver=" + serverName);
     this.onlineServers.put(serverName, sl);
     this.rsAdmins.remove(serverName);
   }
@@ -521,7 +521,7 @@ public class ServerManager {
           }
           sb.append(key);
         }
-        LOG.info("Waiting on regionserver(s) to go down " + sb.toString());
+        LOG.info("Waiting on regionserver(s) " + sb.toString());
         previousLogTime = System.currentTimeMillis();
       }
 
@@ -860,7 +860,7 @@ public class ServerManager {
       if (oldCount != count || lastLogTime + interval < now) {
         lastLogTime = now;
         String msg =
-            "Waiting on RegionServer count=" + count + " to settle; waited="+
+            "Waiting on regionserver count=" + count + "; waited="+
                 slept + "ms, expecting min=" + minToStart + " server(s), max="+ getStrForMax(maxToStart) +
                 " server(s), " + "timeout=" + timeout + "ms, lastChange=" + (lastCountChange - now) + "ms";
         LOG.info(msg);
@@ -883,7 +883,7 @@ public class ServerManager {
     if (isClusterShutdown()) {
       this.master.stop("Cluster shutdown");
     }
-    LOG.info("Finished wait on RegionServer count=" + count + "; waited=" + slept + "ms," +
+    LOG.info("Finished waiting on RegionServer count=" + count + "; waited=" + slept + "ms," +
         " expected min=" + minToStart + " server(s), max=" +  getStrForMax(maxToStart) + " server(s),"+
         " master is "+ (this.master.isStopped() ? "stopped.": "running"));
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
index 2b88fb1..63af398 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/SplitLogManager.java
@@ -53,6 +53,7 @@ import org.apache.hadoop.hbase.monitoring.MonitoredTask;
 import org.apache.hadoop.hbase.monitoring.TaskMonitor;
 import org.apache.hadoop.hbase.util.EnvironmentEdgeManager;
 import org.apache.hadoop.hbase.util.FSUtils;
+import org.apache.hadoop.hbase.util.HasThread;
 import org.apache.hadoop.hbase.wal.AbstractFSWALProvider;
 import org.apache.yetus.audience.InterfaceAudience;
 import org.slf4j.Logger;
@@ -120,7 +121,8 @@ public class SplitLogManager {
       throws IOException {
     this.server = master;
     this.conf = conf;
-    this.choreService = new ChoreService(master.getServerName() + "_splitLogManager_");
+    this.choreService =
+        new ChoreService(((HasThread)master).getName() + ".splitLogManager.");
     if (server.getCoordinatedStateManager() != null) {
       SplitLogManagerCoordination coordination = getSplitLogManagerCoordination();
       Set<String> failedDeletions = Collections.synchronizedSet(new HashSet<String>());

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableStateManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableStateManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableStateManager.java
index b82f47c..021a1a3 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableStateManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/TableStateManager.java
@@ -222,10 +222,11 @@ public class TableStateManager {
     });
     for (Map.Entry<String, TableDescriptor> entry : allDescriptors.entrySet()) {
       String table = entry.getKey();
-      if (table.equals(TableName.META_TABLE_NAME.getNameAsString()))
+      if (table.equals(TableName.META_TABLE_NAME.getNameAsString())) {
         continue;
+      }
       if (!states.containsKey(table)) {
-        LOG.warn("Found table with no state, assuming ENABLED");
+        LOG.warn(table + " has no state, assuming ENABLED");
         MetaTableAccessor.updateTableState(connection, TableName.valueOf(table),
             TableState.State.ENABLED);
       }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
index 4bcd771..1f65230 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentManager.java
@@ -77,6 +77,7 @@ import org.apache.hadoop.hbase.procedure2.ProcedureEvent;
 import org.apache.hadoop.hbase.procedure2.ProcedureExecutor;
 import org.apache.hadoop.hbase.procedure2.ProcedureInMemoryChore;
 import org.apache.hadoop.hbase.procedure2.util.StringUtils;
+import org.apache.hadoop.hbase.util.HasThread;
 import org.apache.hbase.thirdparty.com.google.common.annotations.VisibleForTesting;
 import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProcedureProtos.RegionTransitionState;
@@ -216,7 +217,7 @@ public class AssignmentManager implements ServerListener {
       return;
     }
 
-    LOG.info("Starting assignment manager");
+    LOG.trace("Starting assignment manager");
 
     // Register Server Listener
     master.getServerManager().registerListener(this);
@@ -1182,8 +1183,7 @@ public class AssignmentManager implements ServerListener {
   // ============================================================================================
   public void joinCluster() throws IOException {
     final long startTime = System.currentTimeMillis();
-
-    LOG.info("Joining cluster...Loading hbase:meta content.");
+    LOG.debug("Joining cluster...");
 
     // Scan hbase:meta to build list of existing regions, servers, and assignment
     loadMeta();
@@ -1624,7 +1624,7 @@ public class AssignmentManager implements ServerListener {
   }
 
   private void startAssignmentThread() {
-    assignThread = new Thread("AssignmentThread") {
+    assignThread = new Thread(((HasThread)this.master).getName()) {
       @Override
       public void run() {
         while (isRunning()) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java
index 21f8251..9ef7dce 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/CleanerChore.java
@@ -63,7 +63,7 @@ public abstract class CleanerChore<T extends FileCleanerDelegate> extends Schedu
    * while latter will use only 1 thread for chore to scan dir.
    */
   public static final String CHORE_POOL_SIZE = "hbase.cleaner.scan.dir.concurrent.size";
-  private static final String DEFAULT_CHORE_POOL_SIZE = "0.5";
+  private static final String DEFAULT_CHORE_POOL_SIZE = "0.25";
 
   // It may be waste resources for each cleaner chore own its pool,
   // so let's make pool for all cleaner chores.
@@ -220,9 +220,9 @@ public abstract class CleanerChore<T extends FileCleanerDelegate> extends Schedu
   protected void chore() {
     if (getEnabled()) {
       if (runCleaner()) {
-        LOG.debug("Cleaned old files/dirs under {} successfully",  oldFileDir);
+        LOG.debug("Cleaned all WALs under {}",  oldFileDir);
       } else {
-        LOG.warn("Failed to fully clean old files/dirs under {}", oldFileDir);
+        LOG.warn("WALs outstanding under {}", oldFileDir);
       }
       // After each clean chore, checks if receives reconfigure notification while cleaning
       if (reconfig.compareAndSet(true, false)) {
@@ -382,8 +382,7 @@ public abstract class CleanerChore<T extends FileCleanerDelegate> extends Schedu
     return this.enabled.getAndSet(enabled);
   }
 
-  public boolean getEnabled() {
-    return this.enabled.get();
+  public boolean getEnabled() { return this.enabled.get();
   }
 
   private interface Action<T> {
@@ -405,11 +404,7 @@ public abstract class CleanerChore<T extends FileCleanerDelegate> extends Schedu
 
     @Override
     protected Boolean compute() {
-      if (LOG.isDebugEnabled()) {
-        LOG.debug("CleanerTask " + Thread.currentThread().getId() +
-          " starts cleaning dirs and files under " + dir + " and itself.");
-      }
-
+      LOG.debug("Cleaning under {}", dir);
       List<FileStatus> subDirs;
       List<FileStatus> files;
       try {
@@ -470,13 +465,11 @@ public abstract class CleanerChore<T extends FileCleanerDelegate> extends Schedu
      */
     private boolean deleteAction(Action<Boolean> deletion, String type) {
       boolean deleted;
-      String errorMsg = null;
       try {
         LOG.trace("Start deleting {} under {}", type, dir);
         deleted = deletion.act();
       } catch (IOException ioe) {
-        errorMsg = ioe.getMessage();
-        LOG.warn("Could not delete {} under {}; {}", type, dir, errorMsg);
+        LOG.warn("Could not delete {} under {}; {}", type, dir, ioe);
         deleted = false;
       }
       LOG.trace("Finish deleting {} under {}, deleted=", type, dir, deleted);

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/HFileCleaner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/HFileCleaner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/HFileCleaner.java
index 08640a7..1fbda43 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/HFileCleaner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/HFileCleaner.java
@@ -204,7 +204,7 @@ public class HFileCleaner extends CleanerChore<BaseHFileCleanerDelegate> {
       large.setDaemon(true);
       large.setName(n + "-HFileCleaner.large." + i + "-" + System.currentTimeMillis());
       large.start();
-      LOG.debug("Starting hfile cleaner for large files: {}", large);
+      LOG.debug("Starting for large file={}", large);
       threads.add(large);
     }
 
@@ -219,7 +219,7 @@ public class HFileCleaner extends CleanerChore<BaseHFileCleanerDelegate> {
       small.setDaemon(true);
       small.setName(n + "-HFileCleaner.small." + i + "-" + System.currentTimeMillis());
       small.start();
-      LOG.debug("Starting hfile cleaner for small files: {}", small);
+      LOG.debug("Starting for small files={}", small);
       threads.add(small);
     }
   }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/LogCleaner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/LogCleaner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/LogCleaner.java
index 9beed58..bc27991 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/LogCleaner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/cleaner/LogCleaner.java
@@ -119,7 +119,7 @@ public class LogCleaner extends CleanerChore<BaseLogCleanerDelegate> {
   }
 
   private List<Thread> createOldWalsCleaner(int size) {
-    LOG.info("Creating OldWALs cleaners with size: " + size);
+    LOG.info("Creating OldWALs cleaners with size=" + size);
 
     List<Thread> oldWALsCleaner = new ArrayList<>(size);
     for (int i = 0; i < size; i++) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.java
index ab282d5..2c87745 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/procedure/RSProcedureDispatcher.java
@@ -105,9 +105,8 @@ public class RSProcedureDispatcher
       final Set<RemoteProcedure> remoteProcedures) {
     final int rsVersion = master.getAssignmentManager().getServerVersion(serverName);
     if (rsVersion >= RS_VERSION_WITH_EXEC_PROCS) {
-      LOG.info(String.format(
-        "Using procedure batch rpc execution for serverName=%s version=%s",
-        serverName, rsVersion));
+      LOG.trace("Using procedure batch rpc execution for serverName={} version={}",
+        serverName, rsVersion);
       submitTask(new ExecuteProceduresRemoteCall(serverName, remoteProcedures));
     } else {
       LOG.info(String.format(

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinator.java
index 71ba28e..4184e55 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinator.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureCoordinator.java
@@ -247,7 +247,7 @@ public class ZKProcedureCoordinator implements ProcedureCoordinatorRpcs {
       return false;
     }
 
-    LOG.debug("Starting the controller for procedure member:" + coordName);
+    LOG.debug("Starting controller for procedure member=" + coordName);
     return true;
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureUtil.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureUtil.java
index 9eb3fb3..210bc02 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureUtil.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/procedure/ZKProcedureUtil.java
@@ -269,8 +269,7 @@ public abstract class ZKProcedureUtil
   }
 
   public void clearChildZNodes() throws KeeperException {
-    LOG.info("Clearing all procedure znodes: " + acquiredZnode + " " + reachedZnode + " "
-        + abortZnode);
+    LOG.debug("Clearing all znodes {}, {}, {}", acquiredZnode, reachedZnode, abortZnode);
 
     // If the coordinator was shutdown mid-procedure, then we are going to lose
     // an procedure that was previously started by cleaning out all the previous state. Its much

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChunkCreator.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChunkCreator.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChunkCreator.java
index 44755b4..590ff63 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChunkCreator.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/ChunkCreator.java
@@ -137,8 +137,8 @@ public class ChunkCreator {
       // the pool has run out of maxCount
       if (chunk == null) {
         if (LOG.isTraceEnabled()) {
-          LOG.trace("The chunk pool is full. Reached maxCount= " + this.pool.getMaxCount()
-              + ". Creating chunk onheap.");
+          LOG.trace("Chunk pool full (maxCount={}); creating chunk offheap.",
+              this.pool.getMaxCount());
         }
       }
     }
@@ -163,8 +163,8 @@ public class ChunkCreator {
    */
   Chunk getJumboChunk(CompactingMemStore.IndexType chunkIndexType, int jumboSize) {
     if (jumboSize <= chunkSize) {
-      LOG.warn("Jumbo chunk size " + jumboSize + " must be more than regular chunk size "
-          + chunkSize + ". Converting to regular chunk.");
+      LOG.warn("Jumbo chunk size=" + jumboSize + " must be more than regular chunk size="
+          + chunkSize + "; converting to regular chunk.");
       return getChunk(chunkIndexType,chunkSize);
     }
     // the size of the allocation includes
@@ -407,9 +407,9 @@ public class ChunkCreator {
           MemStoreLAB.CHUNK_POOL_INITIALSIZE_KEY + " must be between 0.0 and 1.0");
     }
     int initialCount = (int) (initialCountPercentage * maxCount);
-    LOG.info("Allocating MemStoreChunkPool with chunk size "
-        + StringUtils.byteDesc(getChunkSize()) + ", max count " + maxCount
-        + ", initial count " + initialCount);
+    LOG.info("Allocating MemStoreChunkPool with chunk size="
+        + StringUtils.byteDesc(getChunkSize()) + ", max count=" + maxCount
+        + ", initial count=" + initialCount);
     return new MemStoreChunkPool(maxCount, initialCount, poolSizePercentage);
   }
 

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
index 1c1ce82..67055c8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegionServer.java
@@ -624,14 +624,14 @@ public class HRegionServer extends HasThread implements
         masterAddressTracker = null;
         clusterStatusTracker = null;
       }
+      this.rpcServices.start();
       // This violates 'no starting stuff in Constructor' but Master depends on the below chore
       // and executor being created and takes a different startup route. Lots of overlap between HRS
       // and M (An M IS A HRS now). Need to refactor so less duplication between M and its super
       // Master expects Constructor to put up web servers. Ugh.
       // class HRS. TODO.
-      this.choreService = new ChoreService(getServerName().toString(), true);
-      this.executorService = new ExecutorService(getServerName().toShortString());
-      this.rpcServices.start();
+      this.choreService = new ChoreService(getName(), true);
+      this.executorService = new ExecutorService(getName());
       putUpWebUI();
     } catch (Throwable t) {
       // Make sure we log the exception. HRegionServer is often started via reflection and the

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
index b6ca599..fd28e05 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/RSRpcServices.java
@@ -100,6 +100,7 @@ import org.apache.hadoop.hbase.ipc.ServerNotRunningYetException;
 import org.apache.hadoop.hbase.ipc.ServerRpcController;
 import org.apache.hadoop.hbase.log.HBaseMarkers;
 import org.apache.hadoop.hbase.master.MasterRpcServices;
+import org.apache.hadoop.hbase.net.Address;
 import org.apache.hadoop.hbase.quotas.ActivePolicyEnforcement;
 import org.apache.hadoop.hbase.quotas.OperationQuota;
 import org.apache.hadoop.hbase.quotas.QuotaUtil;
@@ -1173,7 +1174,11 @@ public class RSRpcServices implements HBaseRPCErrorHandler,
       throw new IllegalArgumentException("Failed resolve of " + initialIsa);
     }
     priority = createPriority();
-    String name = rs.getProcessName() + "/" + initialIsa.toString();
+    String hostname = initialIsa.getHostName();
+    // Using Address means we don't get the IP too. Shorten it more even to just the host name
+    // w/o the domain.
+    String name = rs.getProcessName() + "/" +
+        Address.fromParts(initialIsa.getHostName(), initialIsa.getPort()).toStringWithoutDomain();
     // Set how many times to retry talking to another server over Connection.
     ConnectionUtils.setServerSideHConnectionRetriesConfig(rs.conf, name, LOG);
     rpcServer = createRpcServer(rs, rs.conf, rpcSchedulerFactory, bindAddress, name);

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java
index 86f98da..d2e6d68 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationLogCleaner.java
@@ -83,10 +83,8 @@ public class ReplicationLogCleaner extends BaseLogCleanerDelegate {
       public boolean apply(FileStatus file) {
         String wal = file.getPath().getName();
         boolean logInReplicationQueue = wals.contains(wal);
-        if (LOG.isDebugEnabled()) {
           if (logInReplicationQueue) {
-            LOG.debug("Found log in ZK, keeping: " + wal);
-          }
+            LOG.debug("Found up in ZooKeeper, NOT deleting={}", wal);
         }
         return !logInReplicationQueue && (file.getModificationTime() < readZKTimestamp);
       }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationPeerConfigUpgrader.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationPeerConfigUpgrader.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationPeerConfigUpgrader.java
index ea5509f..d449ba8 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationPeerConfigUpgrader.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/master/ReplicationPeerConfigUpgrader.java
@@ -100,7 +100,7 @@ public class ReplicationPeerConfigUpgrader extends ReplicationStateZKBase {
         // We only need to copy data from tableCFs node to rpc Node the first time hmaster start.
         if (rpc.getTableCFsMap() == null || rpc.getTableCFsMap().isEmpty()) {
           // we copy TableCFs node into PeerNode
-          LOG.info("copy tableCFs into peerNode:" + peerId);
+          LOG.info("Copy table ColumnFamilies into peer=" + peerId);
           ReplicationProtos.TableCF[] tableCFs =
                   ReplicationPeerConfigUtil.parseTableCFs(
                           ZKUtil.getData(this.zookeeper, tableCFsNode));

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
index f191c9d..3a4bf82 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
@@ -756,7 +756,7 @@ public class AccessController implements MasterCoprocessor, RegionCoprocessor,
 
     authorizationEnabled = AccessChecker.isAuthorizationSupported(conf);
     if (!authorizationEnabled) {
-      LOG.warn("The AccessController has been loaded with authorization checks disabled.");
+      LOG.warn("AccessController has been loaded with authorization checks DISABLED!");
     }
 
     shouldCheckExecPermission = conf.getBoolean(AccessControlConstants.EXEC_PERMISSION_CHECKS_KEY,

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
index 2437657..92560cd 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/ZKPermissionWatcher.java
@@ -242,7 +242,7 @@ public class ZKPermissionWatcher extends ZKListener implements Closeable {
 
   private void refreshAuthManager(String entry, byte[] nodeData) throws IOException {
     if (LOG.isDebugEnabled()) {
-      LOG.debug("Updating permissions cache from node "+entry+" with data: "+
+      LOG.debug("Updating permissions cache from {} with data {}", entry,
           Bytes.toStringBinary(nodeData));
     }
     if(AccessControlLists.isNamespaceEntry(entry)) {

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java
index c3f3bd8..49ed11a 100644
--- a/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java
+++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/FSTableDescriptors.java
@@ -289,7 +289,7 @@ public class FSTableDescriptors implements TableDescriptors {
       // add hbase:meta to the response
       tds.put(this.metaTableDescriptor.getTableName().getNameAsString(), metaTableDescriptor);
     } else {
-      LOG.debug("Fetching table descriptors from the filesystem.");
+      LOG.trace("Fetching table descriptors from the filesystem.");
       boolean allvisited = true;
       for (Path d : FSUtils.getTableDirs(fs, rootdir)) {
         TableDescriptor htd = null;
@@ -659,9 +659,9 @@ public class FSTableDescriptors implements TableDescriptors {
       if (sequenceId <= maxSequenceId) {
         boolean success = FSUtils.delete(fs, path, false);
         if (success) {
-          LOG.debug("Deleted table descriptor at " + path);
+          LOG.debug("Deleted " + path);
         } else {
-          LOG.error("Failed to delete descriptor at " + path);
+          LOG.error("Failed to delete table descriptor at " + path);
         }
       }
     }
@@ -713,7 +713,7 @@ public class FSTableDescriptors implements TableDescriptors {
         if (!fs.rename(tempPath, tableInfoDirPath)) {
           throw new IOException("Failed rename of " + tempPath + " to " + tableInfoDirPath);
         }
-        LOG.debug("Wrote descriptor into: " + tableInfoDirPath);
+        LOG.debug("Wrote into " + tableInfoDirPath);
       } catch (IOException ioe) {
         // Presume clash of names or something; go around again.
         LOG.debug("Failed write and/or rename; retrying", ioe);
@@ -784,11 +784,11 @@ public class FSTableDescriptors implements TableDescriptors {
     }
     FileStatus status = getTableInfoPath(fs, tableDir);
     if (status != null) {
-      LOG.debug("Current tableInfoPath = " + status.getPath());
+      LOG.debug("Current path=" + status.getPath());
       if (!forceCreation) {
         if (fs.exists(status.getPath()) && status.getLen() > 0) {
           if (readTableDescriptor(fs, status).equals(htd)) {
-            LOG.debug("TableInfo already exists.. Skipping creation");
+            LOG.trace("TableInfo already exists.. Skipping creation");
             return false;
           }
         }

http://git-wip-us.apache.org/repos/asf/hbase/blob/bac46873/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
----------------------------------------------------------------------
diff --git a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
index 5df31b9..0285a0d 100644
--- a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
+++ b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/zookeeper/RecoverableZooKeeper.java
@@ -568,7 +568,7 @@ public class RecoverableZooKeeper {
                   Bytes.toStringBinary(data));
               throw e;
             }
-            LOG.debug("Node " + path + " already exists");
+            LOG.trace("Node {} already exists", path);
             throw e;
 
           case CONNECTIONLOSS: