You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by da...@apache.org on 2018/08/25 03:09:39 UTC

[02/15] lucene-solr:jira/http2: SOLR-12690: Regularize LoggerFactory declarations

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java
----------------------------------------------------------------------
diff --git a/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java b/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java
index 0121279..f9f1f65 100644
--- a/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java
+++ b/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java
@@ -38,7 +38,7 @@ import org.slf4j.LoggerFactory;
  * @lucene.experimental
  */
 public class BlockDirectory extends FilterDirectory implements ShutdownAwareDirectory {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   
   public static final long BLOCK_SHIFT = Integer.getInteger("solr.hdfs.blockcache.blockshift", 13);
 
@@ -118,11 +118,11 @@ public class BlockDirectory extends FilterDirectory implements ShutdownAwareDire
     }
     this.blockCacheReadEnabled = blockCacheReadEnabled;
     if (!blockCacheReadEnabled) {
-      LOG.info("Block cache on read is disabled");
+      log.info("Block cache on read is disabled");
     }
     this.blockCacheWriteEnabled = blockCacheWriteEnabled;
     if (!blockCacheWriteEnabled) {
-      LOG.info("Block cache on write is disabled");
+      log.info("Block cache on write is disabled");
     }
   }
   
@@ -238,7 +238,7 @@ public class BlockDirectory extends FilterDirectory implements ShutdownAwareDire
   
   @Override
   public void closeOnShutdown() throws IOException {
-    LOG.info("BlockDirectory closing on shutdown");
+    log.info("BlockDirectory closing on shutdown");
     // we are shutting down, no need to clean up cache
     super.close();
   }

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java
----------------------------------------------------------------------
diff --git a/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java b/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java
index 72d48ae..a186715 100644
--- a/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java
+++ b/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java
@@ -41,7 +41,7 @@ import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
 public class HdfsDirectory extends BaseDirectory {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   public static final int DEFAULT_BUFFER_SIZE = 4096;
   
   private static final String LF_EXT = ".lf";
@@ -69,7 +69,7 @@ public class HdfsDirectory extends BaseDirectory {
     if (fileSystem instanceof DistributedFileSystem) {
       // Make sure dfs is not in safe mode
       while (((DistributedFileSystem) fileSystem).setSafeMode(SafeModeAction.SAFEMODE_GET, true)) {
-        LOG.warn("The NameNode is in SafeMode - Solr will wait 5 seconds and try again.");
+        log.warn("The NameNode is in SafeMode - Solr will wait 5 seconds and try again.");
         try {
           Thread.sleep(5000);
         } catch (InterruptedException e) {
@@ -94,7 +94,7 @@ public class HdfsDirectory extends BaseDirectory {
   
   @Override
   public void close() throws IOException {
-    LOG.info("Closing hdfs directory {}", hdfsDirPath);
+    log.info("Closing hdfs directory {}", hdfsDirPath);
     fileSystem.close();
     isOpen = false;
   }
@@ -143,7 +143,7 @@ public class HdfsDirectory extends BaseDirectory {
   @Override
   public void deleteFile(String name) throws IOException {
     Path path = new Path(hdfsDirPath, name);
-    LOG.debug("Deleting {}", path);
+    log.debug("Deleting {}", path);
     getFileSystem().delete(path, false);
   }
   
@@ -197,7 +197,7 @@ public class HdfsDirectory extends BaseDirectory {
   }
   
   public static class HdfsIndexInput extends CustomBufferedIndexInput {
-    private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+    private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
     
     private final Path path;
     private final FSDataInputStream inputStream;
@@ -208,7 +208,7 @@ public class HdfsDirectory extends BaseDirectory {
         int bufferSize) throws IOException {
       super(name, bufferSize);
       this.path = path;
-      LOG.debug("Opening normal index input on {}", path);
+      log.debug("Opening normal index input on {}", path);
       FileStatus fileStatus = fileSystem.getFileStatus(path);
       length = fileStatus.getLen();
       inputStream = fileSystem.open(path, bufferSize);
@@ -227,7 +227,7 @@ public class HdfsDirectory extends BaseDirectory {
     
     @Override
     protected void closeInternal() throws IOException {
-      LOG.debug("Closing normal index input on {}", path);
+      log.debug("Closing normal index input on {}", path);
       if (!clone) {
         inputStream.close();
       }
@@ -248,7 +248,7 @@ public class HdfsDirectory extends BaseDirectory {
   
   @Override
   public void sync(Collection<String> names) throws IOException {
-    LOG.debug("Sync called on {}", Arrays.toString(names.toArray()));
+    log.debug("Sync called on {}", Arrays.toString(names.toArray()));
   }
   
   @Override

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java
----------------------------------------------------------------------
diff --git a/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java b/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java
index 29e2db2..687c505 100644
--- a/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java
+++ b/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java
@@ -57,7 +57,7 @@ import org.slf4j.LoggerFactory;
  * Metrics specific utility functions.
  */
 public class MetricUtils {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   public static final String METRIC_NAME = "metric";
   public static final String VALUE = "value";
@@ -274,7 +274,7 @@ public class MetricUtils {
         convertGauge(n, gauge, propertyFilter, simple, compact, separator, consumer);
       } catch (InternalError ie) {
         if (n.startsWith("memory.") && ie.getMessage().contains("Memory Pool not found")) {
-          LOG.warn("Error converting gauge '" + n + "', possible JDK bug: SOLR-10362", ie);
+          log.warn("Error converting gauge '" + n + "', possible JDK bug: SOLR-10362", ie);
           consumer.accept(n, null);
         } else {
           throw ie;
@@ -577,7 +577,7 @@ public class MetricUtils {
       try {
         beanInfo = Introspector.getBeanInfo(intf, intf.getSuperclass(), Introspector.IGNORE_ALL_BEANINFO);
       } catch (IntrospectionException e) {
-        LOG.warn("Unable to fetch properties of MXBean " + obj.getClass().getName());
+        log.warn("Unable to fetch properties of MXBean " + obj.getClass().getName());
         return;
       }
       for (final PropertyDescriptor desc : beanInfo.getPropertyDescriptors()) {

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java b/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java
index 5346265..a82dd7c 100644
--- a/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java
+++ b/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java
@@ -57,7 +57,7 @@ import static org.apache.solr.common.cloud.Replica.State.DOWN;
 
 public class DeleteReplicaTest extends SolrCloudTestCase {
 
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   @BeforeClass
   public static void setupCluster() throws Exception {
@@ -253,7 +253,7 @@ public class DeleteReplicaTest extends SolrCloudTestCase {
         if (times.incrementAndGet() > 1) {
           return false;
         }
-        LOG.info("Running delete core {}",cd);
+        log.info("Running delete core {}",cd);
 
         try {
           ZkNodeProps m = new ZkNodeProps(
@@ -371,7 +371,7 @@ public class DeleteReplicaTest extends SolrCloudTestCase {
           try {
             cluster.getSolrClient().add(collectionName, new SolrInputDocument("id", String.valueOf(doc++)));
           } catch (Exception e) {
-            LOG.error("Failed on adding document to {}", collectionName, e);
+            log.error("Failed on adding document to {}", collectionName, e);
           }
         }
       });
@@ -389,7 +389,7 @@ public class DeleteReplicaTest extends SolrCloudTestCase {
     try {
       cluster.getSolrClient().waitForState(collectionName, 20, TimeUnit.SECONDS, (liveNodes, collectionState) -> collectionState.getReplicas().size() == 1);
     } catch (TimeoutException e) {
-      LOG.info("Timeout wait for state {}", getCollectionState(collectionName));
+      log.info("Timeout wait for state {}", getCollectionState(collectionName));
       throw e;
     }
 

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java b/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java
index 31947be..54742a9 100644
--- a/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java
+++ b/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java
@@ -54,7 +54,7 @@ import org.slf4j.LoggerFactory;
 @Deprecated
 public class LIROnShardRestartTest extends SolrCloudTestCase {
 
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   @BeforeClass
   public static void setupCluster() throws Exception {
@@ -166,7 +166,7 @@ public class LIROnShardRestartTest extends SolrCloudTestCase {
     } catch (Throwable th) {
       String electionPath = "/collections/allReplicasInLIR/leader_elect/shard1/election/";
       List<String> children = zkClient().getChildren(electionPath, null, true);
-      LOG.info("Election queue {}", children);
+      log.info("Election queue {}", children);
       throw th;
     }
 

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java b/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java
index 3ccd32e..336cef3 100644
--- a/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java
+++ b/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java
@@ -60,7 +60,7 @@ import org.slf4j.LoggerFactory;
 
 public class LIRRollingUpdatesTest extends SolrCloudTestCase {
 
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   private static Map<URI, SocketProxy> proxies;
   private static Map<URI, JettySolrRunner> jettys;
@@ -79,7 +79,7 @@ public class LIRRollingUpdatesTest extends SolrCloudTestCase {
       cluster.stopJettySolrRunner(jetty);//TODO: Can we avoid this restart
       cluster.startJettySolrRunner(jetty);
       proxy.open(jetty.getBaseUrl().toURI());
-      LOG.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
+      log.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
       proxies.put(proxy.getUrl(), proxy);
       jettys.put(proxy.getUrl(), jetty);
     }

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java b/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java
index a4a7bc0..01ad51c 100644
--- a/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java
+++ b/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java
@@ -45,7 +45,7 @@ import org.slf4j.LoggerFactory;
 
 public class LeaderVoteWaitTimeoutTest extends SolrCloudTestCase {
 
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   private static final int NODE_COUNT = 4;
 
   private static Map<JettySolrRunner, SocketProxy> proxies;
@@ -70,7 +70,7 @@ public class LeaderVoteWaitTimeoutTest extends SolrCloudTestCase {
       cluster.stopJettySolrRunner(jetty);//TODO: Can we avoid this restart
       cluster.startJettySolrRunner(jetty);
       proxy.open(jetty.getBaseUrl().toURI());
-      LOG.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
+      log.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
       proxies.put(jetty, proxy);
       jettys.put(proxy.getUrl(), jetty);
     }
@@ -202,7 +202,7 @@ public class LeaderVoteWaitTimeoutTest extends SolrCloudTestCase {
     } catch (Exception e) {
       List<String> children = zkClient().getChildren("/collections/"+collectionName+"/leader_elect/shard1/election",
           null, true);
-      LOG.info("{} election nodes:{}", collectionName, children);
+      log.info("{} election nodes:{}", collectionName, children);
       throw e;
     }
     cluster.getJettySolrRunner(0).start();

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java b/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java
index aac1b9c..6eea5b8 100644
--- a/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java
+++ b/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java
@@ -47,7 +47,7 @@ import org.slf4j.LoggerFactory;
 
 public class TestCloudConsistency extends SolrCloudTestCase {
 
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   private static Map<JettySolrRunner, SocketProxy> proxies;
   private static Map<URI, JettySolrRunner> jettys;
@@ -70,7 +70,7 @@ public class TestCloudConsistency extends SolrCloudTestCase {
       cluster.stopJettySolrRunner(jetty);//TODO: Can we avoid this restart
       cluster.startJettySolrRunner(jetty);
       proxy.open(jetty.getBaseUrl().toURI());
-      LOG.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
+      log.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
       proxies.put(jetty, proxy);
       jettys.put(proxy.getUrl(), jetty);
     }

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java b/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java
index 767fb1b..15625db 100644
--- a/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java
+++ b/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java
@@ -69,7 +69,7 @@ import com.carrotsearch.randomizedtesting.annotations.Repeat;
 @Slow
 public class TestPullReplica extends SolrCloudTestCase {
   
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   
   private String collectionName = null;
   private final static int REPLICATION_TIMEOUT_SECS = 10;
@@ -85,7 +85,7 @@ public class TestPullReplica extends SolrCloudTestCase {
         .addConfig("conf", configset("cloud-minimal"))
         .configure();
     Boolean useLegacyCloud = rarely();
-    LOG.info("Using legacyCloud?: {}", useLegacyCloud);
+    log.info("Using legacyCloud?: {}", useLegacyCloud);
     CollectionAdminRequest.ClusterProp clusterPropRequest = CollectionAdminRequest.setClusterProperty(ZkStateReader.LEGACY_CLOUD, String.valueOf(useLegacyCloud));
     CollectionAdminResponse response = clusterPropRequest.process(cluster.getSolrClient());
     assertEquals(0, response.getStatus());
@@ -107,14 +107,14 @@ public class TestPullReplica extends SolrCloudTestCase {
   public void tearDown() throws Exception {
     for (JettySolrRunner jetty:cluster.getJettySolrRunners()) {
       if (!jetty.isRunning()) {
-        LOG.warn("Jetty {} not running, probably some bad test. Starting it", jetty.getLocalPort());
+        log.warn("Jetty {} not running, probably some bad test. Starting it", jetty.getLocalPort());
         ChaosMonkey.start(jetty);
       }
     }
     if (cluster.getSolrClient().getZkStateReader().getClusterState().getCollectionOrNull(collectionName) != null) {
-      LOG.info("tearDown deleting collection");
+      log.info("tearDown deleting collection");
       CollectionAdminRequest.deleteCollection(collectionName).process(cluster.getSolrClient());
-      LOG.info("Collection deleted");
+      log.info("Collection deleted");
       waitForDeletion(collectionName);
     }
     super.tearDown();
@@ -321,18 +321,18 @@ public class TestPullReplica extends SolrCloudTestCase {
     List<Replica.State> statesSeen = new ArrayList<>(3);
     cluster.getSolrClient().registerCollectionStateWatcher(collectionName, (liveNodes, collectionState) -> {
       Replica r = collectionState.getSlice("shard1").getReplica("core_node2");
-      LOG.info("CollectionStateWatcher state change: {}", r);
+      log.info("CollectionStateWatcher state change: {}", r);
       if (r == null) {
         return false;
       }
       statesSeen.add(r.getState());
-      LOG.info("CollectionStateWatcher saw state: {}", r.getState());
+      log.info("CollectionStateWatcher saw state: {}", r.getState());
       return r.getState() == Replica.State.ACTIVE;
     });
     CollectionAdminRequest.addReplicaToShard(collectionName, "shard1", Replica.Type.PULL).process(cluster.getSolrClient());
     waitForState("Replica not added", collectionName, activeReplicaCount(1, 0, 1));
     zkClient().printLayoutToStdOut();
-    LOG.info("Saw states: " + Arrays.toString(statesSeen.toArray()));
+    log.info("Saw states: " + Arrays.toString(statesSeen.toArray()));
     assertEquals("Expecting DOWN->RECOVERING->ACTIVE but saw: " + Arrays.toString(statesSeen.toArray()), 3, statesSeen.size());
     assertEquals("Expecting DOWN->RECOVERING->ACTIVE but saw: " + Arrays.toString(statesSeen.toArray()), Replica.State.DOWN, statesSeen.get(0));
     assertEquals("Expecting DOWN->RECOVERING->ACTIVE but saw: " + Arrays.toString(statesSeen.toArray()), Replica.State.RECOVERING, statesSeen.get(0));
@@ -557,7 +557,7 @@ public class TestPullReplica extends SolrCloudTestCase {
   private void waitForDeletion(String collection) throws InterruptedException, KeeperException {
     TimeOut t = new TimeOut(10, TimeUnit.SECONDS, TimeSource.NANO_TIME);
     while (cluster.getSolrClient().getZkStateReader().getClusterState().hasCollection(collection)) {
-      LOG.info("Collection not yet deleted");
+      log.info("Collection not yet deleted");
       try {
         Thread.sleep(100);
         if (t.hasTimedOut()) {

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java b/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java
index eb238f6..2c57d33 100644
--- a/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java
+++ b/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java
@@ -58,7 +58,7 @@ public class TestPullReplicaErrorHandling extends SolrCloudTestCase {
   
   private final static int REPLICATION_TIMEOUT_SECS = 10;
   
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   private static Map<URI, SocketProxy> proxies;
   private static Map<URI, JettySolrRunner> jettys;
 
@@ -83,7 +83,7 @@ public class TestPullReplicaErrorHandling extends SolrCloudTestCase {
       cluster.stopJettySolrRunner(jetty);//TODO: Can we avoid this restart
       cluster.startJettySolrRunner(jetty);
       proxy.open(jetty.getBaseUrl().toURI());
-      LOG.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
+      log.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
       proxies.put(proxy.getUrl(), proxy);
       jettys.put(proxy.getUrl(), jetty);
     }
@@ -124,9 +124,9 @@ public class TestPullReplicaErrorHandling extends SolrCloudTestCase {
   @Override
   public void tearDown() throws Exception {
     if (cluster.getSolrClient().getZkStateReader().getClusterState().getCollectionOrNull(collectionName) != null) {
-      LOG.info("tearDown deleting collection");
+      log.info("tearDown deleting collection");
       CollectionAdminRequest.deleteCollection(collectionName).process(cluster.getSolrClient());
-      LOG.info("Collection deleted");
+      log.info("Collection deleted");
       waitForDeletion(collectionName);
     }
     collectionName = null;
@@ -198,7 +198,7 @@ public void testCantConnectToPullReplica() throws Exception {
       }
       assertNumDocs(10, cluster.getSolrClient());
     } finally {
-      LOG.info("Opening leader node");
+      log.info("Opening leader node");
       proxy.reopen();
     }
 //     Back to normal
@@ -304,7 +304,7 @@ public void testCantConnectToPullReplica() throws Exception {
   private void waitForDeletion(String collection) throws InterruptedException, KeeperException {
     TimeOut t = new TimeOut(10, TimeUnit.SECONDS, TimeSource.NANO_TIME);
     while (cluster.getSolrClient().getZkStateReader().getClusterState().hasCollection(collection)) {
-      LOG.info("Collection not yet deleted");
+      log.info("Collection not yet deleted");
       try {
         Thread.sleep(100);
         if (t.hasTimedOut()) {

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java b/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java
index d23fc8d..6888d88 100644
--- a/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java
+++ b/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java
@@ -77,7 +77,7 @@ import org.slf4j.LoggerFactory;
 @Slow
 public class TestTlogReplica extends SolrCloudTestCase {
   
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   
   private String collectionName = null;
   private final static int REPLICATION_TIMEOUT_SECS = 10;
@@ -93,7 +93,7 @@ public class TestTlogReplica extends SolrCloudTestCase {
         .addConfig("conf", configset("cloud-minimal-inplace-updates"))
         .configure();
     Boolean useLegacyCloud = rarely();
-    LOG.info("Using legacyCloud?: {}", useLegacyCloud);
+    log.info("Using legacyCloud?: {}", useLegacyCloud);
     CollectionAdminRequest.ClusterProp clusterPropRequest = CollectionAdminRequest.setClusterProperty(ZkStateReader.LEGACY_CLOUD, String.valueOf(useLegacyCloud));
     CollectionAdminResponse response = clusterPropRequest.process(cluster.getSolrClient());
     assertEquals(0, response.getStatus());
@@ -115,12 +115,12 @@ public class TestTlogReplica extends SolrCloudTestCase {
   public void tearDown() throws Exception {
     for (JettySolrRunner jetty:cluster.getJettySolrRunners()) {
       if (!jetty.isRunning()) {
-        LOG.warn("Jetty {} not running, probably some bad test. Starting it", jetty.getLocalPort());
+        log.warn("Jetty {} not running, probably some bad test. Starting it", jetty.getLocalPort());
         ChaosMonkey.start(jetty);
       }
     }
     if (cluster.getSolrClient().getZkStateReader().getClusterState().getCollectionOrNull(collectionName) != null) {
-      LOG.info("tearDown deleting collection");
+      log.info("tearDown deleting collection");
       CollectionAdminRequest.deleteCollection(collectionName).process(cluster.getSolrClient());
       waitForDeletion(collectionName);
     }
@@ -561,7 +561,7 @@ public class TestTlogReplica extends SolrCloudTestCase {
       if ((Integer)((NamedList<Object>)response.get("responseHeader")).get(UpdateRequest.REPFACT) >= 2) {
         break;
       }
-      LOG.info("Min RF not achieved yet. retrying");
+      log.info("Min RF not achieved yet. retrying");
     }
     checkRTG(3,7, cluster.getJettySolrRunners());
     DirectUpdateHandler2.commitOnClose = false;
@@ -603,7 +603,7 @@ public class TestTlogReplica extends SolrCloudTestCase {
       if ((Integer)((NamedList<Object>)response.get("responseHeader")).get(UpdateRequest.REPFACT) >= 2) {
         break;
       }
-      LOG.info("Min RF not achieved yet. retrying");
+      log.info("Min RF not achieved yet. retrying");
     }
     new UpdateRequest()
         .add(sdoc("id", "9"))

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java
index 76bea430..727ff64 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java
@@ -57,7 +57,7 @@ import org.slf4j.LoggerFactory;
  * Implementation based on {@link org.apache.solr.cloud.ZkDistributedQueue}
  */
 public class GenericDistributedQueue implements DistributedQueue {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   static final String PREFIX = "qn-";
 
@@ -252,7 +252,7 @@ public class GenericDistributedQueue implements DistributedQueue {
             try {
               stateManager.removeData(ops.get(j).getPath(), -1);
             } catch (NoSuchElementException e2) {
-              LOG.debug("Can not remove node which is not exist : " + ops.get(j).getPath());
+              log.debug("Can not remove node which is not exist : " + ops.get(j).getPath());
             }
           }
         }
@@ -419,7 +419,7 @@ public class GenericDistributedQueue implements DistributedQueue {
         for (String childName : childNames) {
           // Check format
           if (!childName.regionMatches(0, PREFIX, 0, PREFIX.length())) {
-            LOG.debug("Found child node with improper name: " + childName);
+            log.debug("Found child node with improper name: " + childName);
             continue;
           }
           orderedChildren.add(childName);

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java
index 63dd5bf..1f0b6cf 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java
@@ -105,7 +105,7 @@ import static org.apache.solr.cloud.api.collections.OverseerCollectionMessageHan
  * Simulated {@link SolrCloudManager}.
  */
 public class SimCloudManager implements SolrCloudManager {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   private final SimDistribStateManager stateManager;
   private final SimClusterStateProvider clusterStateProvider;
@@ -395,7 +395,7 @@ public class SimCloudManager implements SolrCloudManager {
     String nodeId = (String)values.get(ImplicitSnitch.NODE);
     nodeStateProvider.simSetNodeValues(nodeId, values);
     clusterStateProvider.simAddNode(nodeId);
-    LOG.trace("-- added node " + nodeId);
+    log.trace("-- added node " + nodeId);
     // initialize history handler if this is the first node
     if (metricsHistoryHandler == null && liveNodesSet.size() == 1) {
       metricsHandler = new MetricsHandler(metricManager);
@@ -428,7 +428,7 @@ public class SimCloudManager implements SolrCloudManager {
         metricsHandler = null;
       }
     }
-    LOG.trace("-- removed node " + nodeId);
+    log.trace("-- removed node " + nodeId);
   }
 
   /**
@@ -517,7 +517,7 @@ public class SimCloudManager implements SolrCloudManager {
    * @param killNodeId optional nodeId to kill. If null then don't kill any node, just restart the thread
    */
   public void simRestartOverseer(String killNodeId) throws Exception {
-    LOG.info("=== Restarting OverseerTriggerThread and clearing object cache...");
+    log.info("=== Restarting OverseerTriggerThread and clearing object cache...");
     triggerThread.interrupt();
     IOUtils.closeQuietly(triggerThread);
     if (killNodeId != null) {
@@ -648,7 +648,7 @@ public class SimCloudManager implements SolrCloudManager {
     // pay the penalty for remote request, at least 5 ms
     timeSource.sleep(5);
 
-    LOG.trace("--- got SolrRequest: " + req.getMethod() + " " + req.getPath() +
+    log.trace("--- got SolrRequest: " + req.getMethod() + " " + req.getPath() +
         (req.getParams() != null ? " " + req.getParams().toQueryString() : ""));
     if (req.getPath() != null) {
       if (req.getPath().startsWith("/admin/autoscaling") ||
@@ -674,7 +674,7 @@ public class SimCloudManager implements SolrCloudManager {
           ByteArrayOutputStream baos = new ByteArrayOutputStream();
           cw.write(baos);
           String payload = baos.toString("UTF-8");
-          LOG.trace("-- payload: {}", payload);
+          log.trace("-- payload: {}", payload);
           queryRequest.setContentStreams(Collections.singletonList(new ContentStreamBase.StringStream(payload)));
         }
         queryRequest.getContext().put("httpMethod", req.getMethod().toString());
@@ -698,12 +698,12 @@ public class SimCloudManager implements SolrCloudManager {
           }
         }
         if (queryResponse.getException() != null) {
-          LOG.debug("-- exception handling request", queryResponse.getException());
+          log.debug("-- exception handling request", queryResponse.getException());
           throw new IOException(queryResponse.getException());
         }
         SolrResponse rsp = new SolrResponseBase();
         rsp.setResponse(queryResponse.getValues());
-        LOG.trace("-- response: {}", rsp);
+        log.trace("-- response: {}", rsp);
         return rsp;
       }
     }
@@ -736,7 +736,7 @@ public class SimCloudManager implements SolrCloudManager {
       if (action == null) {
         throw new SolrException(SolrException.ErrorCode.BAD_REQUEST, "Unknown action: " + a);
       }
-      LOG.trace("Invoking Collection Action :{} with params {}", action.toLower(), req.getParams().toQueryString());
+      log.trace("Invoking Collection Action :{} with params {}", action.toLower(), req.getParams().toQueryString());
       NamedList results = new NamedList();
       rsp.setResponse(results);
       incrementCount(action.name());

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java
index 741a868..17b56d7 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java
@@ -113,7 +113,7 @@ import static org.apache.solr.common.params.CommonParams.NAME;
  *   </ul>
  */
 public class SimClusterStateProvider implements ClusterStateProvider {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   public static final long DEFAULT_DOC_SIZE_BYTES = 500;
 
@@ -333,7 +333,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
     }
     // pick first
     overseerLeader = liveNodes.iterator().next();
-    LOG.debug("--- new Overseer leader: " + overseerLeader);
+    log.debug("--- new Overseer leader: " + overseerLeader);
     // record it in ZK
     Map<String, Object> id = new HashMap<>();
     id.put("id", cloudManager.getTimeSource().getTimeNs() +
@@ -341,7 +341,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
     try {
       cloudManager.getDistribStateManager().makePath(path, Utils.toJSON(id), CreateMode.EPHEMERAL, false);
     } catch (Exception e) {
-      LOG.warn("Exception saving overseer leader id", e);
+      log.warn("Exception saving overseer leader id", e);
     }
   }
 
@@ -513,7 +513,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
           "", true, "INDEX.sizeInBytes");
       // at this point nuke our cached DocCollection state
       collectionsStatesRef.set(null);
-      LOG.trace("-- simAddReplica {}", replicaInfo);
+      log.trace("-- simAddReplica {}", replicaInfo);
       if (runLeaderElection) {
         simRunLeaderElection(Collections.singleton(replicaInfo.getCollection()), true);
       }
@@ -552,7 +552,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
             }
             cloudManager.getSimNodeStateProvider().simSetNodeValue(nodeId, ImplicitSnitch.DISK, disk + 1);
           }
-          LOG.trace("-- simRemoveReplica {}", ri);
+          log.trace("-- simRemoveReplica {}", ri);
           simRunLeaderElection(Collections.singleton(ri.getCollection()), true);
           return;
         }
@@ -612,14 +612,14 @@ public class SimClusterStateProvider implements ClusterStateProvider {
       }
       dc.getSlices().forEach(s -> {
         if (s.getLeader() != null) {
-          LOG.debug("-- already has leader {} / {}", dc.getName(), s.getName());
+          log.debug("-- already has leader {} / {}", dc.getName(), s.getName());
           return;
         }
         if (s.getReplicas().isEmpty()) {
-          LOG.debug("-- no replicas in {} / {}", dc.getName(), s.getName());
+          log.debug("-- no replicas in {} / {}", dc.getName(), s.getName());
           return;
         }
-        LOG.debug("-- submit leader election for {} / {}", dc.getName(), s.getName());
+        log.debug("-- submit leader election for {} / {}", dc.getName(), s.getName());
         cloudManager.submit(() -> {
           simRunLeaderElection(dc.getName(), s, saveClusterState);
           return true;
@@ -632,9 +632,9 @@ public class SimClusterStateProvider implements ClusterStateProvider {
     AtomicBoolean stateChanged = new AtomicBoolean(Boolean.FALSE);
     Replica leader = s.getLeader();
     if (leader == null || !liveNodes.contains(leader.getNodeName())) {
-      LOG.debug("Running leader election for {} / {}", collection, s.getName());
+      log.debug("Running leader election for {} / {}", collection, s.getName());
       if (s.getReplicas().isEmpty()) { // no replicas - punt
-        LOG.debug("-- no replicas in {} / {}", collection, s.getName());
+        log.debug("-- no replicas in {} / {}", collection, s.getName());
         return;
       }
       ActionThrottle lt = getThrottle(collection, s.getName());
@@ -651,14 +651,14 @@ public class SimClusterStateProvider implements ClusterStateProvider {
           synchronized (ri) {
             if (r.isActive(liveNodes.get())) {
               if (ri.getVariables().get(ZkStateReader.LEADER_PROP) != null) {
-                LOG.trace("-- found existing leader {} / {}: {}, {}", collection, s.getName(), ri, r);
+                log.trace("-- found existing leader {} / {}: {}, {}", collection, s.getName(), ri, r);
                 alreadyHasLeader.set(true);
                 return;
               } else {
                 active.add(ri);
               }
             } else { // if it's on a node that is not live mark it down
-              LOG.trace("-- replica not active on live nodes: {}, {}", liveNodes.get(), r);
+              log.trace("-- replica not active on live nodes: {}, {}", liveNodes.get(), r);
               if (!liveNodes.contains(r.getNodeName())) {
                 ri.getVariables().put(ZkStateReader.STATE_PROP, Replica.State.DOWN.toString());
                 ri.getVariables().remove(ZkStateReader.LEADER_PROP);
@@ -668,12 +668,12 @@ public class SimClusterStateProvider implements ClusterStateProvider {
           }
         });
         if (alreadyHasLeader.get()) {
-          LOG.debug("-- already has leader {} / {}: {}", collection, s.getName(), s);
+          log.debug("-- already has leader {} / {}: {}", collection, s.getName(), s);
           return;
         }
         if (active.isEmpty()) {
-          LOG.warn("Can't find any active replicas for {} / {}: {}", collection, s.getName(), s);
-          LOG.debug("-- liveNodes: {}", liveNodes.get());
+          log.warn("Can't find any active replicas for {} / {}: {}", collection, s.getName(), s);
+          log.debug("-- liveNodes: {}", liveNodes.get());
           return;
         }
         // pick first active one
@@ -685,7 +685,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
           }
         }
         if (ri == null) {
-          LOG.warn("-- can't find any suitable replica type for {} / {}: {}", collection, s.getName(), s);
+          log.warn("-- can't find any suitable replica type for {} / {}: {}", collection, s.getName(), s);
           return;
         }
         // now mark the leader election throttle
@@ -695,10 +695,10 @@ public class SimClusterStateProvider implements ClusterStateProvider {
           ri.getVariables().put(ZkStateReader.LEADER_PROP, "true");
         }
         stateChanged.set(true);
-        LOG.debug("-- elected new leader for " + collection + " / " + s.getName() + ": " + ri.getName());
+        log.debug("-- elected new leader for " + collection + " / " + s.getName() + ": " + ri.getName());
       }
     } else {
-      LOG.debug("-- already has leader for {} / {}", collection, s.getName());
+      log.debug("-- already has leader for {} / {}", collection, s.getName());
     }
     if (stateChanged.get() || saveState) {
       collectionsStatesRef.set(null);
@@ -751,7 +751,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
 
     ZkWriteCommand cmd = new ClusterStateMutator(cloudManager).createCollection(clusterState, props);
     if (cmd.noop) {
-      LOG.warn("Collection {} already exists. exit", collectionName);
+      log.warn("Collection {} already exists. exit", collectionName);
       results.add("success", "no-op");
       return;
     }
@@ -906,7 +906,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
       saveClusterState.set(true);
       results.add("success", "");
     } catch (Exception e) {
-      LOG.warn("Exception", e);
+      log.warn("Exception", e);
     } finally {
       lock.unlock();
     }
@@ -973,7 +973,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
     String newSolrCoreName = Assign.buildSolrCoreName(stateManager, coll, slice.getName(), replica.getType());
     String coreNodeName = Assign.assignCoreNodeName(stateManager, coll);
     ReplicaInfo newReplica = new ReplicaInfo(coreNodeName, newSolrCoreName, collection, slice.getName(), replica.getType(), targetNode, null);
-    LOG.debug("-- new replica: " + newReplica);
+    log.debug("-- new replica: " + newReplica);
     // xxx should run leader election here already?
     simAddReplica(targetNode, newReplica, false);
     // this will trigger leader election
@@ -1276,14 +1276,14 @@ public class SimClusterStateProvider implements ClusterStateProvider {
           // NOTE: we don't use getProperty because it uses PROPERTY_PROP_PREFIX
           Replica leader = s.getLeader();
           if (leader == null) {
-            LOG.debug("-- no leader in " + s);
+            log.debug("-- no leader in " + s);
             continue;
           }
           cloudManager.getMetricManager().registry(createRegistryName(collection, s.getName(), leader)).counter("UPDATE./update.requests").inc();
           ReplicaInfo ri = getReplicaInfo(leader);
           Number numDocs = (Number)ri.getVariable("SEARCHER.searcher.numDocs");
           if (numDocs == null || numDocs.intValue() <= 0) {
-            LOG.debug("-- attempting to delete nonexistent doc " + id + " from " + s.getLeader());
+            log.debug("-- attempting to delete nonexistent doc " + id + " from " + s.getLeader());
             continue;
           }
           modified = true;
@@ -1314,7 +1314,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
           for (Slice s : coll.getSlices()) {
             Replica leader = s.getLeader();
             if (leader == null) {
-              LOG.debug("-- no leader in " + s);
+              log.debug("-- no leader in " + s);
               continue;
             }
 
@@ -1348,7 +1348,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
           Slice s = router.getTargetSlice(id, null, null, req.getParams(), coll);
           Replica leader = s.getLeader();
           if (leader == null) {
-            LOG.debug("-- no leader in " + s);
+            log.debug("-- no leader in " + s);
             continue;
           }
           cloudManager.getMetricManager().registry(createRegistryName(collection, s.getName(), leader)).counter("UPDATE./update.requests").inc();

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java
index 7a8dbbb..1e99ff2 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java
@@ -68,7 +68,7 @@ import org.slf4j.LoggerFactory;
  * invoked.
  */
 public class SimDistribStateManager implements DistribStateManager {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   public static final class Node {
     ReentrantLock dataLock = new ReentrantLock();

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java
index e9616f0..7168ce9 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java
@@ -51,7 +51,7 @@ import org.slf4j.LoggerFactory;
  * exposed anywhere.
  */
 public class SimDistributedQueueFactory implements DistributedQueueFactory {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   Map<String, SimDistributedQueue> queues = new ConcurrentHashMap<>();
 
@@ -190,7 +190,7 @@ public class SimDistributedQueueFactory implements DistributedQueueFactory {
       try {
         queue.offer(new Pair(String.format(Locale.ROOT, "qn-%010d", seq), data));
         seq++;
-        LOG.trace("=== offer " + System.nanoTime());
+        log.trace("=== offer " + System.nanoTime());
         changed.signalAll();
       } finally {
         updateLock.unlock();

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java
index 7a346ea..9673fa7 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java
@@ -47,7 +47,7 @@ import org.slf4j.LoggerFactory;
  * to setup core-level metrics use {@link SimClusterStateProvider#simSetCollectionValue(String, String, Object, boolean, boolean)}.
  */
 public class SimNodeStateProvider implements NodeStateProvider {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   private final Map<String, Map<String, Object>> nodeValues = new ConcurrentHashMap<>();
   private final SimClusterStateProvider clusterStateProvider;
@@ -164,7 +164,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
    * @param node node id
    */
   public void simRemoveNodeValues(String node) throws InterruptedException {
-    LOG.debug("--removing value for " + node);
+    log.debug("--removing value for " + node);
     lock.lockInterruptibly();
     try {
       Map<String, Object> values = nodeValues.remove(node);
@@ -187,7 +187,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
     try {
       AtomicBoolean updateRoles = new AtomicBoolean(false);
       myNodes.forEach(n -> {
-        LOG.debug("- removing dead node values: " + n);
+        log.debug("- removing dead node values: " + n);
         Map<String, Object> vals = nodeValues.remove(n);
         if (vals.containsKey("nodeRole")) {
           updateRoles.set(true);
@@ -253,7 +253,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
     for (String tag : tags) {
       String[] parts = tag.split(":");
       if (parts.length < 3 || !parts[0].equals("metrics")) {
-        LOG.warn("Invalid metrics: tag: " + tag);
+        log.warn("Invalid metrics: tag: " + tag);
         continue;
       }
       if (!parts[1].startsWith("solr.core.")) {
@@ -263,7 +263,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
       Matcher m = REGISTRY_PATTERN.matcher(parts[1]);
 
       if (!m.matches()) {
-        LOG.warn("Invalid registry name: " + parts[1]);
+        log.warn("Invalid registry name: " + parts[1]);
         continue;
       }
       String collection = m.group(1);
@@ -291,7 +291,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
 
   @Override
   public Map<String, Object> getNodeValues(String node, Collection<String> tags) {
-    LOG.trace("-- requested values for " + node + ": " + tags);
+    log.trace("-- requested values for " + node + ": " + tags);
     if (!liveNodesSet.contains(node)) {
       throw new RuntimeException("non-live node " + node);
     }

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java
index e395985..2cdc456 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java
@@ -54,7 +54,7 @@ import org.slf4j.LoggerFactory;
  * This test compares the cluster state of a real cluster and a simulated one.
  */
 public class TestClusterStateProvider extends SolrCloudTestCase {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   private static int NODE_COUNT = 3;
   private static boolean simulated;
@@ -72,7 +72,7 @@ public class TestClusterStateProvider extends SolrCloudTestCase {
   @BeforeClass
   public static void setupCluster() throws Exception {
     simulated = random().nextBoolean();
-    LOG.info("####### Using simulated components? " + simulated);
+    log.info("####### Using simulated components? " + simulated);
 
     configureCluster(NODE_COUNT)
         .addConfig("conf", configset("cloud-minimal"))

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java
index 14f7416..74d9bb1 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java
@@ -47,7 +47,7 @@ import org.slf4j.LoggerFactory;
  * This test compares a ZK-based {@link DistribStateManager} to the simulated one.
  */
 public class TestDistribStateManager extends SolrTestCaseJ4 {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   private DistribStateManager stateManager;
   private ZkTestServer zkTestServer;
@@ -80,7 +80,7 @@ public class TestDistribStateManager extends SolrTestCaseJ4 {
       solrZkClient = new SolrZkClient(zkTestServer.getZkHost(), 30000);
       stateManager = new ZkDistribStateManager(solrZkClient);
     }
-    LOG.info("Using " + stateManager.getClass().getName());
+    log.info("Using " + stateManager.getClass().getName());
   }
 
   private DistribStateManager createDistribStateManager() {

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java b/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java
index 652f3f0..362940a 100644
--- a/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java
+++ b/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java
@@ -42,7 +42,7 @@ import org.junit.Test;
 
 public class CurrencyRangeFacetCloudTest extends SolrCloudTestCase {
 
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   
   private static final String COLLECTION = MethodHandles.lookup().lookupClass().getName();
   private static final String CONF = COLLECTION + "_configSet";

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java b/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java
index 05c25cf..846977c 100644
--- a/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java
+++ b/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java
@@ -53,7 +53,7 @@ import org.junit.BeforeClass;
  */
 public class RangeFacetCloudTest extends SolrCloudTestCase {
 
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   
   private static final String COLLECTION = MethodHandles.lookup().lookupClass().getName();
   private static final String CONF = COLLECTION + "_configSet";

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java b/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java
index fa0505e..ccd02eb 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java
@@ -43,7 +43,7 @@ import static java.util.stream.Collectors.toList;
  * lazily.
  */
 public class AutoScalingConfig implements MapWriter {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   private final Map<String, Object> jsonMap;
   private final boolean empty;
@@ -81,7 +81,7 @@ public class AutoScalingConfig implements MapWriter {
           TriggerEventProcessorStage stage = TriggerEventProcessorStage.valueOf(String.valueOf(stageName).toUpperCase(Locale.ROOT));
           stages.add(stage);
         } catch (Exception e) {
-          LOG.warn("Invalid stage name '" + name + "' in listener config, skipping: " + properties);
+          log.warn("Invalid stage name '" + name + "' in listener config, skipping: " + properties);
         }
       }
       listenerClass = (String)this.properties.get(AutoScalingParams.CLASS);

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java b/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java
index 525728e..210e324 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java
@@ -69,7 +69,7 @@ import static org.apache.solr.client.solrj.cloud.autoscaling.Variable.Type.WITH_
  *
  */
 public class Policy implements MapWriter {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   public static final String POLICY = "policy";
   public static final String EACH = "#EACH";
@@ -281,7 +281,7 @@ public class Policy implements MapWriter {
             return p.compare(r1, r2, false);
           });
         } catch (Exception e) {
-          LOG.error("Exception! prefs = {}, recent r1 = {}, r2 = {}, matrix = {}",
+          log.error("Exception! prefs = {}, recent r1 = {}, r2 = {}, matrix = {}",
               clusterPreferences,
               lastComparison[0],
               lastComparison[1],
@@ -498,9 +498,9 @@ public class Policy implements MapWriter {
       this.nodeStateProvider = cloudManager.getNodeStateProvider();
       try {
         state = cloudManager.getClusterStateProvider().getClusterState();
-        LOG.trace("-- session created with cluster state: {}", state);
+        log.trace("-- session created with cluster state: {}", state);
       } catch (Exception e) {
-        LOG.trace("-- session created, can't obtain cluster state", e);
+        log.trace("-- session created, can't obtain cluster state", e);
       }
       this.znodeVersion = state != null ? state.getZNodeVersion() : -1;
       this.nodes = new ArrayList<>(cloudManager.getClusterStateProvider().getLiveNodes());

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java
index e08f85f..d538b18 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java
@@ -73,7 +73,7 @@ import org.slf4j.LoggerFactory;
  */
 public class HttpClientUtil {
   
-  private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   
   public static final int DEFAULT_CONNECT_TIMEOUT = 60000;
   public static final int DEFAULT_SO_TIMEOUT = 600000;
@@ -147,7 +147,7 @@ public class HttpClientUtil {
     // Configure the HttpClientBuilder if user has specified the factory type.
     String factoryClassName = System.getProperty(SYS_PROP_HTTP_CLIENT_BUILDER_FACTORY);
     if (factoryClassName != null) {
-      logger.debug ("Using " + factoryClassName);
+      log.debug ("Using " + factoryClassName);
       try {
         HttpClientBuilderFactory factory = (HttpClientBuilderFactory)Class.forName(factoryClassName).newInstance();
         httpClientBuilder = factory.getHttpClientBuilder(Optional.of(SolrHttpClientBuilder.create()));
@@ -176,7 +176,7 @@ public class HttpClientUtil {
           try {
             interceptor.process(request, context);
           } catch (Exception e) {
-            logger.error("", e);
+            log.error("", e);
           }
         }
       });
@@ -234,7 +234,7 @@ public class HttpClientUtil {
       } else {
         sslConnectionSocketFactory = new SSLConnectionSocketFactory(SSLContexts.createSystemDefault(),
                                                                     NoopHostnameVerifier.INSTANCE);
-        logger.debug(HttpClientUtil.SYS_PROP_CHECK_PEER_NAME + "is false, hostname checks disabled.");
+        log.debug(HttpClientUtil.SYS_PROP_CHECK_PEER_NAME + "is false, hostname checks disabled.");
       }
       builder.register("https", sslConnectionSocketFactory);
 
@@ -268,8 +268,8 @@ public class HttpClientUtil {
 
   public static CloseableHttpClient createClient(final SolrParams params, PoolingHttpClientConnectionManager cm, boolean sharedConnectionManager, HttpRequestExecutor httpRequestExecutor)  {
     final ModifiableSolrParams config = new ModifiableSolrParams(params);
-    if (logger.isDebugEnabled()) {
-      logger.debug("Creating new http client, config:" + config);
+    if (log.isDebugEnabled()) {
+      log.debug("Creating new http client, config:" + config);
     }
 
     cm.setMaxTotal(params.getInt(HttpClientUtil.PROP_MAX_CONNECTIONS, 10000));

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java
index 1bcf96b..afa2ef6 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java
@@ -50,7 +50,7 @@ import org.slf4j.LoggerFactory;
 public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
   
   public static final String LOGIN_CONFIG_PROP = "java.security.auth.login.config";
-  private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
   
   private static Configuration jaasConfig = new SolrJaasConfiguration();
 
@@ -85,7 +85,7 @@ public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
       String configValue = System.getProperty(LOGIN_CONFIG_PROP);
 
       if (configValue != null) {
-        logger.info("Setting up SPNego auth with config: " + configValue);
+        log.info("Setting up SPNego auth with config: " + configValue);
         final String useSubjectCredsProp = "javax.security.auth.useSubjectCredsOnly";
         String useSubjectCredsVal = System.getProperty(useSubjectCredsProp);
 
@@ -97,7 +97,7 @@ public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
         else if (!useSubjectCredsVal.toLowerCase(Locale.ROOT).equals("false")) {
           // Don't overwrite the prop value if it's already been written to something else,
           // but log because it is likely the Credentials won't be loaded correctly.
-          logger.warn("System Property: " + useSubjectCredsProp + " set to: " + useSubjectCredsVal
+          log.warn("System Property: " + useSubjectCredsProp + " set to: " + useSubjectCredsVal
               + " not false.  SPNego authentication may not be successful.");
         }
 
@@ -139,7 +139,7 @@ public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
         HttpClientUtil.addRequestInterceptor(bufferedEntityInterceptor);
       }
     } else {
-      logger.warn("{} is configured without specifying system property '{}'",
+      log.warn("{} is configured without specifying system property '{}'",
           getClass().getName(), LOGIN_CONFIG_PROP);
     }
 
@@ -176,11 +176,11 @@ public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
     public AppConfigurationEntry[] getAppConfigurationEntry(String appName) {
       if (baseConfig == null) return null;
 
-      logger.debug("Login prop: "+System.getProperty(LOGIN_CONFIG_PROP));
+      log.debug("Login prop: "+System.getProperty(LOGIN_CONFIG_PROP));
 
       String clientAppName = System.getProperty("solr.kerberos.jaas.appname", "Client");
       if (initiateAppNames.contains(appName)) {
-        logger.debug("Using AppConfigurationEntry for appName '"+clientAppName+"' instead of: " + appName);
+        log.debug("Using AppConfigurationEntry for appName '"+clientAppName+"' instead of: " + appName);
         return baseConfig.getAppConfigurationEntry(clientAppName);
       }
       return baseConfig.getAppConfigurationEntry(appName);

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java
index dc229bf..b29ea09 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java
@@ -43,7 +43,7 @@ import org.slf4j.LoggerFactory;
  */
 public class CommitStream extends TupleStream implements Expressible {
   private static final long serialVersionUID = 1L;
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   // Part of expression / passed in
   private String collection;
@@ -252,7 +252,7 @@ public class CommitStream extends TupleStream implements Expressible {
     try {
       clientCache.getCloudSolrClient(zkHost).commit(collection, waitFlush, waitSearcher, softCommit);
     } catch (SolrServerException | IOException e) {
-      LOG.warn(String.format(Locale.ROOT, "Unable to commit documents to collection '%s' due to unexpected error.", collection), e);
+      log.warn(String.format(Locale.ROOT, "Unable to commit documents to collection '%s' due to unexpected error.", collection), e);
       String className = e.getClass().getName();
       String message = e.getMessage();
       throw new IOException(String.format(Locale.ROOT,"Unexpected error when committing documents to collection %s- %s:%s", collection, className, message));

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java
index a4f528b..9d02ec2 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java
@@ -60,7 +60,7 @@ public class DaemonStream extends TupleStream implements Expressible {
   private Map<String, DaemonStream> daemons;
   private boolean terminate;
   private boolean closed = false;
-  private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   public DaemonStream(StreamExpression expression, StreamFactory factory) throws IOException{
 
@@ -329,14 +329,14 @@ public class DaemonStream extends TupleStream implements Expressible {
             }
           } catch (IOException e) {
             exception = e;
-            logger.error("Error in DaemonStream:" + id, e);
+            log.error("Error in DaemonStream:" + id, e);
             ++errors;
             if (errors > 100) {
-              logger.error("Too many consectutive errors. Stopping DaemonStream:" + id);
+              log.error("Too many consectutive errors. Stopping DaemonStream:" + id);
               break OUTER;
             }
           } catch (Throwable t) {
-            logger.error("Fatal Error in DaemonStream:" + id, t);
+            log.error("Fatal Error in DaemonStream:" + id, t);
             //For anything other then IOException break out of the loop and shutdown the thread.
             break OUTER;
           } finally {
@@ -345,7 +345,7 @@ public class DaemonStream extends TupleStream implements Expressible {
             } catch (IOException e1) {
               if (exception == null) {
                 exception = e1;
-                logger.error("Error in DaemonStream:" + id, e1);
+                log.error("Error in DaemonStream:" + id, e1);
                 break OUTER;
               }
             }
@@ -357,7 +357,7 @@ public class DaemonStream extends TupleStream implements Expressible {
           try {
             Thread.sleep(sleepMillis);
           } catch (InterruptedException e) {
-            logger.error("Error in DaemonStream:" + id, e);
+            log.error("Error in DaemonStream:" + id, e);
             break OUTER;
           }
         }
@@ -370,7 +370,7 @@ public class DaemonStream extends TupleStream implements Expressible {
         try {
           queue.put(tuple);
         } catch (InterruptedException e) {
-          logger.error("Error in DaemonStream:"+id, e);
+          log.error("Error in DaemonStream:"+id, e);
         }
       }
       setStopTime(new Date().getTime());

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java
index 2a1b491..bee208c 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java
@@ -55,7 +55,7 @@ import static org.apache.solr.common.params.CommonParams.ID;
 
 public class ExecutorStream extends TupleStream implements Expressible {
 
-  private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   private TupleStream stream;
 
@@ -148,7 +148,7 @@ public class ExecutorStream extends TupleStream implements Expressible {
     try {
       executorService.awaitTermination(Long.MAX_VALUE, TimeUnit.SECONDS);
     } catch(InterruptedException e) {
-      logger.error("Interrupted while waiting for termination", e);
+      log.error("Interrupted while waiting for termination", e);
     }
   }
 
@@ -214,12 +214,12 @@ public class ExecutorStream extends TupleStream implements Expressible {
           }
         }
       } catch (Exception e) {
-        logger.error("Executor Error: id="+id+" expr_s="+expr, e);
+        log.error("Executor Error: id="+id+" expr_s="+expr, e);
       } finally {
         try {
           stream.close();
         } catch (Exception e1) {
-          logger.error("Executor Error", e1);
+          log.error("Executor Error", e1);
         }
       }
     }

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java
index a08f838..c00de10 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java
@@ -50,7 +50,7 @@ import static org.apache.solr.common.params.CommonParams.VERSION_FIELD;
  * @since 6.0.0
  */
 public class UpdateStream extends TupleStream implements Expressible {
-  private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   public static String BATCH_INDEXED_FIELD_NAME = "batchIndexed"; // field name in summary tuple for #docs updated in batch
   private String collection;
@@ -281,7 +281,7 @@ public class UpdateStream extends TupleStream implements Expressible {
         }
       }
     }
-    LOG.debug("Tuple [{}] was converted into SolrInputDocument [{}].", tuple, doc);
+    log.debug("Tuple [{}] was converted into SolrInputDocument [{}].", tuple, doc);
     
     return doc;
   }
@@ -300,7 +300,7 @@ public class UpdateStream extends TupleStream implements Expressible {
     try {
       cloudSolrClient.add(collection, documentBatch);
     } catch (SolrServerException | IOException e) {
-      LOG.warn("Unable to add documents to collection due to unexpected error.", e);
+      log.warn("Unable to add documents to collection due to unexpected error.", e);
       String className = e.getClass().getName();
       String message = e.getMessage();
       throw new IOException(String.format(Locale.ROOT,"Unexpected error when adding documents to collection %s- %s:%s", collection, className, message));

http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java b/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java
index 015793c..bf4a1ed 100644
--- a/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java
+++ b/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java
@@ -36,7 +36,7 @@ import java.util.regex.Pattern;
  */
 public class ZkConfigManager {
 
-  private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+  private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
 
   /** ZkNode where named configs are stored */
   public static final String CONFIGS_ZKNODE = "/configs";
@@ -139,7 +139,7 @@ public class ZkConfigManager {
         List<String> children = zkClient.getChildren(fromZkPath + "/" + file, null, true);
         if (children.size() == 0) {
           final String toZkFilePath = toZkPath + "/" + file;
-          logger.info("Copying zk node {} to {}",
+          log.info("Copying zk node {} to {}",
               fromZkPath + "/" + file, toZkFilePath);
           byte[] data = zkClient.getData(fromZkPath + "/" + file, null, null, true);
           zkClient.makePath(toZkFilePath, data, true);