You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by er...@apache.org on 2018/08/23 02:50:36 UTC
[2/5] lucene-solr:master: SOLR-12690: Regularize LoggerFactory
declarations
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java
----------------------------------------------------------------------
diff --git a/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java b/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java
index 0121279..f9f1f65 100644
--- a/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java
+++ b/solr/core/src/java/org/apache/solr/store/blockcache/BlockDirectory.java
@@ -38,7 +38,7 @@ import org.slf4j.LoggerFactory;
* @lucene.experimental
*/
public class BlockDirectory extends FilterDirectory implements ShutdownAwareDirectory {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public static final long BLOCK_SHIFT = Integer.getInteger("solr.hdfs.blockcache.blockshift", 13);
@@ -118,11 +118,11 @@ public class BlockDirectory extends FilterDirectory implements ShutdownAwareDire
}
this.blockCacheReadEnabled = blockCacheReadEnabled;
if (!blockCacheReadEnabled) {
- LOG.info("Block cache on read is disabled");
+ log.info("Block cache on read is disabled");
}
this.blockCacheWriteEnabled = blockCacheWriteEnabled;
if (!blockCacheWriteEnabled) {
- LOG.info("Block cache on write is disabled");
+ log.info("Block cache on write is disabled");
}
}
@@ -238,7 +238,7 @@ public class BlockDirectory extends FilterDirectory implements ShutdownAwareDire
@Override
public void closeOnShutdown() throws IOException {
- LOG.info("BlockDirectory closing on shutdown");
+ log.info("BlockDirectory closing on shutdown");
// we are shutting down, no need to clean up cache
super.close();
}
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java
----------------------------------------------------------------------
diff --git a/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java b/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java
index 72d48ae..a186715 100644
--- a/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java
+++ b/solr/core/src/java/org/apache/solr/store/hdfs/HdfsDirectory.java
@@ -41,7 +41,7 @@ import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
public class HdfsDirectory extends BaseDirectory {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public static final int DEFAULT_BUFFER_SIZE = 4096;
private static final String LF_EXT = ".lf";
@@ -69,7 +69,7 @@ public class HdfsDirectory extends BaseDirectory {
if (fileSystem instanceof DistributedFileSystem) {
// Make sure dfs is not in safe mode
while (((DistributedFileSystem) fileSystem).setSafeMode(SafeModeAction.SAFEMODE_GET, true)) {
- LOG.warn("The NameNode is in SafeMode - Solr will wait 5 seconds and try again.");
+ log.warn("The NameNode is in SafeMode - Solr will wait 5 seconds and try again.");
try {
Thread.sleep(5000);
} catch (InterruptedException e) {
@@ -94,7 +94,7 @@ public class HdfsDirectory extends BaseDirectory {
@Override
public void close() throws IOException {
- LOG.info("Closing hdfs directory {}", hdfsDirPath);
+ log.info("Closing hdfs directory {}", hdfsDirPath);
fileSystem.close();
isOpen = false;
}
@@ -143,7 +143,7 @@ public class HdfsDirectory extends BaseDirectory {
@Override
public void deleteFile(String name) throws IOException {
Path path = new Path(hdfsDirPath, name);
- LOG.debug("Deleting {}", path);
+ log.debug("Deleting {}", path);
getFileSystem().delete(path, false);
}
@@ -197,7 +197,7 @@ public class HdfsDirectory extends BaseDirectory {
}
public static class HdfsIndexInput extends CustomBufferedIndexInput {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private final Path path;
private final FSDataInputStream inputStream;
@@ -208,7 +208,7 @@ public class HdfsDirectory extends BaseDirectory {
int bufferSize) throws IOException {
super(name, bufferSize);
this.path = path;
- LOG.debug("Opening normal index input on {}", path);
+ log.debug("Opening normal index input on {}", path);
FileStatus fileStatus = fileSystem.getFileStatus(path);
length = fileStatus.getLen();
inputStream = fileSystem.open(path, bufferSize);
@@ -227,7 +227,7 @@ public class HdfsDirectory extends BaseDirectory {
@Override
protected void closeInternal() throws IOException {
- LOG.debug("Closing normal index input on {}", path);
+ log.debug("Closing normal index input on {}", path);
if (!clone) {
inputStream.close();
}
@@ -248,7 +248,7 @@ public class HdfsDirectory extends BaseDirectory {
@Override
public void sync(Collection<String> names) throws IOException {
- LOG.debug("Sync called on {}", Arrays.toString(names.toArray()));
+ log.debug("Sync called on {}", Arrays.toString(names.toArray()));
}
@Override
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java
----------------------------------------------------------------------
diff --git a/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java b/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java
index 29e2db2..687c505 100644
--- a/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java
+++ b/solr/core/src/java/org/apache/solr/util/stats/MetricUtils.java
@@ -57,7 +57,7 @@ import org.slf4j.LoggerFactory;
* Metrics specific utility functions.
*/
public class MetricUtils {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public static final String METRIC_NAME = "metric";
public static final String VALUE = "value";
@@ -274,7 +274,7 @@ public class MetricUtils {
convertGauge(n, gauge, propertyFilter, simple, compact, separator, consumer);
} catch (InternalError ie) {
if (n.startsWith("memory.") && ie.getMessage().contains("Memory Pool not found")) {
- LOG.warn("Error converting gauge '" + n + "', possible JDK bug: SOLR-10362", ie);
+ log.warn("Error converting gauge '" + n + "', possible JDK bug: SOLR-10362", ie);
consumer.accept(n, null);
} else {
throw ie;
@@ -577,7 +577,7 @@ public class MetricUtils {
try {
beanInfo = Introspector.getBeanInfo(intf, intf.getSuperclass(), Introspector.IGNORE_ALL_BEANINFO);
} catch (IntrospectionException e) {
- LOG.warn("Unable to fetch properties of MXBean " + obj.getClass().getName());
+ log.warn("Unable to fetch properties of MXBean " + obj.getClass().getName());
return;
}
for (final PropertyDescriptor desc : beanInfo.getPropertyDescriptors()) {
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java b/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java
index 5346265..a82dd7c 100644
--- a/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java
+++ b/solr/core/src/test/org/apache/solr/cloud/DeleteReplicaTest.java
@@ -57,7 +57,7 @@ import static org.apache.solr.common.cloud.Replica.State.DOWN;
public class DeleteReplicaTest extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
@BeforeClass
public static void setupCluster() throws Exception {
@@ -253,7 +253,7 @@ public class DeleteReplicaTest extends SolrCloudTestCase {
if (times.incrementAndGet() > 1) {
return false;
}
- LOG.info("Running delete core {}",cd);
+ log.info("Running delete core {}",cd);
try {
ZkNodeProps m = new ZkNodeProps(
@@ -371,7 +371,7 @@ public class DeleteReplicaTest extends SolrCloudTestCase {
try {
cluster.getSolrClient().add(collectionName, new SolrInputDocument("id", String.valueOf(doc++)));
} catch (Exception e) {
- LOG.error("Failed on adding document to {}", collectionName, e);
+ log.error("Failed on adding document to {}", collectionName, e);
}
}
});
@@ -389,7 +389,7 @@ public class DeleteReplicaTest extends SolrCloudTestCase {
try {
cluster.getSolrClient().waitForState(collectionName, 20, TimeUnit.SECONDS, (liveNodes, collectionState) -> collectionState.getReplicas().size() == 1);
} catch (TimeoutException e) {
- LOG.info("Timeout wait for state {}", getCollectionState(collectionName));
+ log.info("Timeout wait for state {}", getCollectionState(collectionName));
throw e;
}
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java b/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java
index 31947be..54742a9 100644
--- a/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java
+++ b/solr/core/src/test/org/apache/solr/cloud/LIROnShardRestartTest.java
@@ -54,7 +54,7 @@ import org.slf4j.LoggerFactory;
@Deprecated
public class LIROnShardRestartTest extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
@BeforeClass
public static void setupCluster() throws Exception {
@@ -166,7 +166,7 @@ public class LIROnShardRestartTest extends SolrCloudTestCase {
} catch (Throwable th) {
String electionPath = "/collections/allReplicasInLIR/leader_elect/shard1/election/";
List<String> children = zkClient().getChildren(electionPath, null, true);
- LOG.info("Election queue {}", children);
+ log.info("Election queue {}", children);
throw th;
}
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java b/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java
index 3ccd32e..336cef3 100644
--- a/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java
+++ b/solr/core/src/test/org/apache/solr/cloud/LIRRollingUpdatesTest.java
@@ -60,7 +60,7 @@ import org.slf4j.LoggerFactory;
public class LIRRollingUpdatesTest extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private static Map<URI, SocketProxy> proxies;
private static Map<URI, JettySolrRunner> jettys;
@@ -79,7 +79,7 @@ public class LIRRollingUpdatesTest extends SolrCloudTestCase {
cluster.stopJettySolrRunner(jetty);//TODO: Can we avoid this restart
cluster.startJettySolrRunner(jetty);
proxy.open(jetty.getBaseUrl().toURI());
- LOG.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
+ log.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
proxies.put(proxy.getUrl(), proxy);
jettys.put(proxy.getUrl(), jetty);
}
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java b/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java
index a4a7bc0..01ad51c 100644
--- a/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java
+++ b/solr/core/src/test/org/apache/solr/cloud/LeaderVoteWaitTimeoutTest.java
@@ -45,7 +45,7 @@ import org.slf4j.LoggerFactory;
public class LeaderVoteWaitTimeoutTest extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private static final int NODE_COUNT = 4;
private static Map<JettySolrRunner, SocketProxy> proxies;
@@ -70,7 +70,7 @@ public class LeaderVoteWaitTimeoutTest extends SolrCloudTestCase {
cluster.stopJettySolrRunner(jetty);//TODO: Can we avoid this restart
cluster.startJettySolrRunner(jetty);
proxy.open(jetty.getBaseUrl().toURI());
- LOG.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
+ log.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
proxies.put(jetty, proxy);
jettys.put(proxy.getUrl(), jetty);
}
@@ -202,7 +202,7 @@ public class LeaderVoteWaitTimeoutTest extends SolrCloudTestCase {
} catch (Exception e) {
List<String> children = zkClient().getChildren("/collections/"+collectionName+"/leader_elect/shard1/election",
null, true);
- LOG.info("{} election nodes:{}", collectionName, children);
+ log.info("{} election nodes:{}", collectionName, children);
throw e;
}
cluster.getJettySolrRunner(0).start();
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java b/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java
index aac1b9c..6eea5b8 100644
--- a/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java
+++ b/solr/core/src/test/org/apache/solr/cloud/TestCloudConsistency.java
@@ -47,7 +47,7 @@ import org.slf4j.LoggerFactory;
public class TestCloudConsistency extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private static Map<JettySolrRunner, SocketProxy> proxies;
private static Map<URI, JettySolrRunner> jettys;
@@ -70,7 +70,7 @@ public class TestCloudConsistency extends SolrCloudTestCase {
cluster.stopJettySolrRunner(jetty);//TODO: Can we avoid this restart
cluster.startJettySolrRunner(jetty);
proxy.open(jetty.getBaseUrl().toURI());
- LOG.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
+ log.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
proxies.put(jetty, proxy);
jettys.put(proxy.getUrl(), jetty);
}
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java b/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java
index 767fb1b..15625db 100644
--- a/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java
+++ b/solr/core/src/test/org/apache/solr/cloud/TestPullReplica.java
@@ -69,7 +69,7 @@ import com.carrotsearch.randomizedtesting.annotations.Repeat;
@Slow
public class TestPullReplica extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private String collectionName = null;
private final static int REPLICATION_TIMEOUT_SECS = 10;
@@ -85,7 +85,7 @@ public class TestPullReplica extends SolrCloudTestCase {
.addConfig("conf", configset("cloud-minimal"))
.configure();
Boolean useLegacyCloud = rarely();
- LOG.info("Using legacyCloud?: {}", useLegacyCloud);
+ log.info("Using legacyCloud?: {}", useLegacyCloud);
CollectionAdminRequest.ClusterProp clusterPropRequest = CollectionAdminRequest.setClusterProperty(ZkStateReader.LEGACY_CLOUD, String.valueOf(useLegacyCloud));
CollectionAdminResponse response = clusterPropRequest.process(cluster.getSolrClient());
assertEquals(0, response.getStatus());
@@ -107,14 +107,14 @@ public class TestPullReplica extends SolrCloudTestCase {
public void tearDown() throws Exception {
for (JettySolrRunner jetty:cluster.getJettySolrRunners()) {
if (!jetty.isRunning()) {
- LOG.warn("Jetty {} not running, probably some bad test. Starting it", jetty.getLocalPort());
+ log.warn("Jetty {} not running, probably some bad test. Starting it", jetty.getLocalPort());
ChaosMonkey.start(jetty);
}
}
if (cluster.getSolrClient().getZkStateReader().getClusterState().getCollectionOrNull(collectionName) != null) {
- LOG.info("tearDown deleting collection");
+ log.info("tearDown deleting collection");
CollectionAdminRequest.deleteCollection(collectionName).process(cluster.getSolrClient());
- LOG.info("Collection deleted");
+ log.info("Collection deleted");
waitForDeletion(collectionName);
}
super.tearDown();
@@ -321,18 +321,18 @@ public class TestPullReplica extends SolrCloudTestCase {
List<Replica.State> statesSeen = new ArrayList<>(3);
cluster.getSolrClient().registerCollectionStateWatcher(collectionName, (liveNodes, collectionState) -> {
Replica r = collectionState.getSlice("shard1").getReplica("core_node2");
- LOG.info("CollectionStateWatcher state change: {}", r);
+ log.info("CollectionStateWatcher state change: {}", r);
if (r == null) {
return false;
}
statesSeen.add(r.getState());
- LOG.info("CollectionStateWatcher saw state: {}", r.getState());
+ log.info("CollectionStateWatcher saw state: {}", r.getState());
return r.getState() == Replica.State.ACTIVE;
});
CollectionAdminRequest.addReplicaToShard(collectionName, "shard1", Replica.Type.PULL).process(cluster.getSolrClient());
waitForState("Replica not added", collectionName, activeReplicaCount(1, 0, 1));
zkClient().printLayoutToStdOut();
- LOG.info("Saw states: " + Arrays.toString(statesSeen.toArray()));
+ log.info("Saw states: " + Arrays.toString(statesSeen.toArray()));
assertEquals("Expecting DOWN->RECOVERING->ACTIVE but saw: " + Arrays.toString(statesSeen.toArray()), 3, statesSeen.size());
assertEquals("Expecting DOWN->RECOVERING->ACTIVE but saw: " + Arrays.toString(statesSeen.toArray()), Replica.State.DOWN, statesSeen.get(0));
assertEquals("Expecting DOWN->RECOVERING->ACTIVE but saw: " + Arrays.toString(statesSeen.toArray()), Replica.State.RECOVERING, statesSeen.get(0));
@@ -557,7 +557,7 @@ public class TestPullReplica extends SolrCloudTestCase {
private void waitForDeletion(String collection) throws InterruptedException, KeeperException {
TimeOut t = new TimeOut(10, TimeUnit.SECONDS, TimeSource.NANO_TIME);
while (cluster.getSolrClient().getZkStateReader().getClusterState().hasCollection(collection)) {
- LOG.info("Collection not yet deleted");
+ log.info("Collection not yet deleted");
try {
Thread.sleep(100);
if (t.hasTimedOut()) {
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java b/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java
index eb238f6..2c57d33 100644
--- a/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java
+++ b/solr/core/src/test/org/apache/solr/cloud/TestPullReplicaErrorHandling.java
@@ -58,7 +58,7 @@ public class TestPullReplicaErrorHandling extends SolrCloudTestCase {
private final static int REPLICATION_TIMEOUT_SECS = 10;
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private static Map<URI, SocketProxy> proxies;
private static Map<URI, JettySolrRunner> jettys;
@@ -83,7 +83,7 @@ public class TestPullReplicaErrorHandling extends SolrCloudTestCase {
cluster.stopJettySolrRunner(jetty);//TODO: Can we avoid this restart
cluster.startJettySolrRunner(jetty);
proxy.open(jetty.getBaseUrl().toURI());
- LOG.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
+ log.info("Adding proxy for URL: " + jetty.getBaseUrl() + ". Proxy: " + proxy.getUrl());
proxies.put(proxy.getUrl(), proxy);
jettys.put(proxy.getUrl(), jetty);
}
@@ -124,9 +124,9 @@ public class TestPullReplicaErrorHandling extends SolrCloudTestCase {
@Override
public void tearDown() throws Exception {
if (cluster.getSolrClient().getZkStateReader().getClusterState().getCollectionOrNull(collectionName) != null) {
- LOG.info("tearDown deleting collection");
+ log.info("tearDown deleting collection");
CollectionAdminRequest.deleteCollection(collectionName).process(cluster.getSolrClient());
- LOG.info("Collection deleted");
+ log.info("Collection deleted");
waitForDeletion(collectionName);
}
collectionName = null;
@@ -198,7 +198,7 @@ public void testCantConnectToPullReplica() throws Exception {
}
assertNumDocs(10, cluster.getSolrClient());
} finally {
- LOG.info("Opening leader node");
+ log.info("Opening leader node");
proxy.reopen();
}
// Back to normal
@@ -304,7 +304,7 @@ public void testCantConnectToPullReplica() throws Exception {
private void waitForDeletion(String collection) throws InterruptedException, KeeperException {
TimeOut t = new TimeOut(10, TimeUnit.SECONDS, TimeSource.NANO_TIME);
while (cluster.getSolrClient().getZkStateReader().getClusterState().hasCollection(collection)) {
- LOG.info("Collection not yet deleted");
+ log.info("Collection not yet deleted");
try {
Thread.sleep(100);
if (t.hasTimedOut()) {
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java b/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java
index d23fc8d..6888d88 100644
--- a/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java
+++ b/solr/core/src/test/org/apache/solr/cloud/TestTlogReplica.java
@@ -77,7 +77,7 @@ import org.slf4j.LoggerFactory;
@Slow
public class TestTlogReplica extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private String collectionName = null;
private final static int REPLICATION_TIMEOUT_SECS = 10;
@@ -93,7 +93,7 @@ public class TestTlogReplica extends SolrCloudTestCase {
.addConfig("conf", configset("cloud-minimal-inplace-updates"))
.configure();
Boolean useLegacyCloud = rarely();
- LOG.info("Using legacyCloud?: {}", useLegacyCloud);
+ log.info("Using legacyCloud?: {}", useLegacyCloud);
CollectionAdminRequest.ClusterProp clusterPropRequest = CollectionAdminRequest.setClusterProperty(ZkStateReader.LEGACY_CLOUD, String.valueOf(useLegacyCloud));
CollectionAdminResponse response = clusterPropRequest.process(cluster.getSolrClient());
assertEquals(0, response.getStatus());
@@ -115,12 +115,12 @@ public class TestTlogReplica extends SolrCloudTestCase {
public void tearDown() throws Exception {
for (JettySolrRunner jetty:cluster.getJettySolrRunners()) {
if (!jetty.isRunning()) {
- LOG.warn("Jetty {} not running, probably some bad test. Starting it", jetty.getLocalPort());
+ log.warn("Jetty {} not running, probably some bad test. Starting it", jetty.getLocalPort());
ChaosMonkey.start(jetty);
}
}
if (cluster.getSolrClient().getZkStateReader().getClusterState().getCollectionOrNull(collectionName) != null) {
- LOG.info("tearDown deleting collection");
+ log.info("tearDown deleting collection");
CollectionAdminRequest.deleteCollection(collectionName).process(cluster.getSolrClient());
waitForDeletion(collectionName);
}
@@ -561,7 +561,7 @@ public class TestTlogReplica extends SolrCloudTestCase {
if ((Integer)((NamedList<Object>)response.get("responseHeader")).get(UpdateRequest.REPFACT) >= 2) {
break;
}
- LOG.info("Min RF not achieved yet. retrying");
+ log.info("Min RF not achieved yet. retrying");
}
checkRTG(3,7, cluster.getJettySolrRunners());
DirectUpdateHandler2.commitOnClose = false;
@@ -603,7 +603,7 @@ public class TestTlogReplica extends SolrCloudTestCase {
if ((Integer)((NamedList<Object>)response.get("responseHeader")).get(UpdateRequest.REPFACT) >= 2) {
break;
}
- LOG.info("Min RF not achieved yet. retrying");
+ log.info("Min RF not achieved yet. retrying");
}
new UpdateRequest()
.add(sdoc("id", "9"))
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java
index 76bea430..727ff64 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/GenericDistributedQueue.java
@@ -57,7 +57,7 @@ import org.slf4j.LoggerFactory;
* Implementation based on {@link org.apache.solr.cloud.ZkDistributedQueue}
*/
public class GenericDistributedQueue implements DistributedQueue {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
static final String PREFIX = "qn-";
@@ -252,7 +252,7 @@ public class GenericDistributedQueue implements DistributedQueue {
try {
stateManager.removeData(ops.get(j).getPath(), -1);
} catch (NoSuchElementException e2) {
- LOG.debug("Can not remove node which is not exist : " + ops.get(j).getPath());
+ log.debug("Can not remove node which is not exist : " + ops.get(j).getPath());
}
}
}
@@ -419,7 +419,7 @@ public class GenericDistributedQueue implements DistributedQueue {
for (String childName : childNames) {
// Check format
if (!childName.regionMatches(0, PREFIX, 0, PREFIX.length())) {
- LOG.debug("Found child node with improper name: " + childName);
+ log.debug("Found child node with improper name: " + childName);
continue;
}
orderedChildren.add(childName);
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java
index 63dd5bf..1f0b6cf 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimCloudManager.java
@@ -105,7 +105,7 @@ import static org.apache.solr.cloud.api.collections.OverseerCollectionMessageHan
* Simulated {@link SolrCloudManager}.
*/
public class SimCloudManager implements SolrCloudManager {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private final SimDistribStateManager stateManager;
private final SimClusterStateProvider clusterStateProvider;
@@ -395,7 +395,7 @@ public class SimCloudManager implements SolrCloudManager {
String nodeId = (String)values.get(ImplicitSnitch.NODE);
nodeStateProvider.simSetNodeValues(nodeId, values);
clusterStateProvider.simAddNode(nodeId);
- LOG.trace("-- added node " + nodeId);
+ log.trace("-- added node " + nodeId);
// initialize history handler if this is the first node
if (metricsHistoryHandler == null && liveNodesSet.size() == 1) {
metricsHandler = new MetricsHandler(metricManager);
@@ -428,7 +428,7 @@ public class SimCloudManager implements SolrCloudManager {
metricsHandler = null;
}
}
- LOG.trace("-- removed node " + nodeId);
+ log.trace("-- removed node " + nodeId);
}
/**
@@ -517,7 +517,7 @@ public class SimCloudManager implements SolrCloudManager {
* @param killNodeId optional nodeId to kill. If null then don't kill any node, just restart the thread
*/
public void simRestartOverseer(String killNodeId) throws Exception {
- LOG.info("=== Restarting OverseerTriggerThread and clearing object cache...");
+ log.info("=== Restarting OverseerTriggerThread and clearing object cache...");
triggerThread.interrupt();
IOUtils.closeQuietly(triggerThread);
if (killNodeId != null) {
@@ -648,7 +648,7 @@ public class SimCloudManager implements SolrCloudManager {
// pay the penalty for remote request, at least 5 ms
timeSource.sleep(5);
- LOG.trace("--- got SolrRequest: " + req.getMethod() + " " + req.getPath() +
+ log.trace("--- got SolrRequest: " + req.getMethod() + " " + req.getPath() +
(req.getParams() != null ? " " + req.getParams().toQueryString() : ""));
if (req.getPath() != null) {
if (req.getPath().startsWith("/admin/autoscaling") ||
@@ -674,7 +674,7 @@ public class SimCloudManager implements SolrCloudManager {
ByteArrayOutputStream baos = new ByteArrayOutputStream();
cw.write(baos);
String payload = baos.toString("UTF-8");
- LOG.trace("-- payload: {}", payload);
+ log.trace("-- payload: {}", payload);
queryRequest.setContentStreams(Collections.singletonList(new ContentStreamBase.StringStream(payload)));
}
queryRequest.getContext().put("httpMethod", req.getMethod().toString());
@@ -698,12 +698,12 @@ public class SimCloudManager implements SolrCloudManager {
}
}
if (queryResponse.getException() != null) {
- LOG.debug("-- exception handling request", queryResponse.getException());
+ log.debug("-- exception handling request", queryResponse.getException());
throw new IOException(queryResponse.getException());
}
SolrResponse rsp = new SolrResponseBase();
rsp.setResponse(queryResponse.getValues());
- LOG.trace("-- response: {}", rsp);
+ log.trace("-- response: {}", rsp);
return rsp;
}
}
@@ -736,7 +736,7 @@ public class SimCloudManager implements SolrCloudManager {
if (action == null) {
throw new SolrException(SolrException.ErrorCode.BAD_REQUEST, "Unknown action: " + a);
}
- LOG.trace("Invoking Collection Action :{} with params {}", action.toLower(), req.getParams().toQueryString());
+ log.trace("Invoking Collection Action :{} with params {}", action.toLower(), req.getParams().toQueryString());
NamedList results = new NamedList();
rsp.setResponse(results);
incrementCount(action.name());
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java
index 741a868..17b56d7 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimClusterStateProvider.java
@@ -113,7 +113,7 @@ import static org.apache.solr.common.params.CommonParams.NAME;
* </ul>
*/
public class SimClusterStateProvider implements ClusterStateProvider {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public static final long DEFAULT_DOC_SIZE_BYTES = 500;
@@ -333,7 +333,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
}
// pick first
overseerLeader = liveNodes.iterator().next();
- LOG.debug("--- new Overseer leader: " + overseerLeader);
+ log.debug("--- new Overseer leader: " + overseerLeader);
// record it in ZK
Map<String, Object> id = new HashMap<>();
id.put("id", cloudManager.getTimeSource().getTimeNs() +
@@ -341,7 +341,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
try {
cloudManager.getDistribStateManager().makePath(path, Utils.toJSON(id), CreateMode.EPHEMERAL, false);
} catch (Exception e) {
- LOG.warn("Exception saving overseer leader id", e);
+ log.warn("Exception saving overseer leader id", e);
}
}
@@ -513,7 +513,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
"", true, "INDEX.sizeInBytes");
// at this point nuke our cached DocCollection state
collectionsStatesRef.set(null);
- LOG.trace("-- simAddReplica {}", replicaInfo);
+ log.trace("-- simAddReplica {}", replicaInfo);
if (runLeaderElection) {
simRunLeaderElection(Collections.singleton(replicaInfo.getCollection()), true);
}
@@ -552,7 +552,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
}
cloudManager.getSimNodeStateProvider().simSetNodeValue(nodeId, ImplicitSnitch.DISK, disk + 1);
}
- LOG.trace("-- simRemoveReplica {}", ri);
+ log.trace("-- simRemoveReplica {}", ri);
simRunLeaderElection(Collections.singleton(ri.getCollection()), true);
return;
}
@@ -612,14 +612,14 @@ public class SimClusterStateProvider implements ClusterStateProvider {
}
dc.getSlices().forEach(s -> {
if (s.getLeader() != null) {
- LOG.debug("-- already has leader {} / {}", dc.getName(), s.getName());
+ log.debug("-- already has leader {} / {}", dc.getName(), s.getName());
return;
}
if (s.getReplicas().isEmpty()) {
- LOG.debug("-- no replicas in {} / {}", dc.getName(), s.getName());
+ log.debug("-- no replicas in {} / {}", dc.getName(), s.getName());
return;
}
- LOG.debug("-- submit leader election for {} / {}", dc.getName(), s.getName());
+ log.debug("-- submit leader election for {} / {}", dc.getName(), s.getName());
cloudManager.submit(() -> {
simRunLeaderElection(dc.getName(), s, saveClusterState);
return true;
@@ -632,9 +632,9 @@ public class SimClusterStateProvider implements ClusterStateProvider {
AtomicBoolean stateChanged = new AtomicBoolean(Boolean.FALSE);
Replica leader = s.getLeader();
if (leader == null || !liveNodes.contains(leader.getNodeName())) {
- LOG.debug("Running leader election for {} / {}", collection, s.getName());
+ log.debug("Running leader election for {} / {}", collection, s.getName());
if (s.getReplicas().isEmpty()) { // no replicas - punt
- LOG.debug("-- no replicas in {} / {}", collection, s.getName());
+ log.debug("-- no replicas in {} / {}", collection, s.getName());
return;
}
ActionThrottle lt = getThrottle(collection, s.getName());
@@ -651,14 +651,14 @@ public class SimClusterStateProvider implements ClusterStateProvider {
synchronized (ri) {
if (r.isActive(liveNodes.get())) {
if (ri.getVariables().get(ZkStateReader.LEADER_PROP) != null) {
- LOG.trace("-- found existing leader {} / {}: {}, {}", collection, s.getName(), ri, r);
+ log.trace("-- found existing leader {} / {}: {}, {}", collection, s.getName(), ri, r);
alreadyHasLeader.set(true);
return;
} else {
active.add(ri);
}
} else { // if it's on a node that is not live mark it down
- LOG.trace("-- replica not active on live nodes: {}, {}", liveNodes.get(), r);
+ log.trace("-- replica not active on live nodes: {}, {}", liveNodes.get(), r);
if (!liveNodes.contains(r.getNodeName())) {
ri.getVariables().put(ZkStateReader.STATE_PROP, Replica.State.DOWN.toString());
ri.getVariables().remove(ZkStateReader.LEADER_PROP);
@@ -668,12 +668,12 @@ public class SimClusterStateProvider implements ClusterStateProvider {
}
});
if (alreadyHasLeader.get()) {
- LOG.debug("-- already has leader {} / {}: {}", collection, s.getName(), s);
+ log.debug("-- already has leader {} / {}: {}", collection, s.getName(), s);
return;
}
if (active.isEmpty()) {
- LOG.warn("Can't find any active replicas for {} / {}: {}", collection, s.getName(), s);
- LOG.debug("-- liveNodes: {}", liveNodes.get());
+ log.warn("Can't find any active replicas for {} / {}: {}", collection, s.getName(), s);
+ log.debug("-- liveNodes: {}", liveNodes.get());
return;
}
// pick first active one
@@ -685,7 +685,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
}
}
if (ri == null) {
- LOG.warn("-- can't find any suitable replica type for {} / {}: {}", collection, s.getName(), s);
+ log.warn("-- can't find any suitable replica type for {} / {}: {}", collection, s.getName(), s);
return;
}
// now mark the leader election throttle
@@ -695,10 +695,10 @@ public class SimClusterStateProvider implements ClusterStateProvider {
ri.getVariables().put(ZkStateReader.LEADER_PROP, "true");
}
stateChanged.set(true);
- LOG.debug("-- elected new leader for " + collection + " / " + s.getName() + ": " + ri.getName());
+ log.debug("-- elected new leader for " + collection + " / " + s.getName() + ": " + ri.getName());
}
} else {
- LOG.debug("-- already has leader for {} / {}", collection, s.getName());
+ log.debug("-- already has leader for {} / {}", collection, s.getName());
}
if (stateChanged.get() || saveState) {
collectionsStatesRef.set(null);
@@ -751,7 +751,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
ZkWriteCommand cmd = new ClusterStateMutator(cloudManager).createCollection(clusterState, props);
if (cmd.noop) {
- LOG.warn("Collection {} already exists. exit", collectionName);
+ log.warn("Collection {} already exists. exit", collectionName);
results.add("success", "no-op");
return;
}
@@ -906,7 +906,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
saveClusterState.set(true);
results.add("success", "");
} catch (Exception e) {
- LOG.warn("Exception", e);
+ log.warn("Exception", e);
} finally {
lock.unlock();
}
@@ -973,7 +973,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
String newSolrCoreName = Assign.buildSolrCoreName(stateManager, coll, slice.getName(), replica.getType());
String coreNodeName = Assign.assignCoreNodeName(stateManager, coll);
ReplicaInfo newReplica = new ReplicaInfo(coreNodeName, newSolrCoreName, collection, slice.getName(), replica.getType(), targetNode, null);
- LOG.debug("-- new replica: " + newReplica);
+ log.debug("-- new replica: " + newReplica);
// xxx should run leader election here already?
simAddReplica(targetNode, newReplica, false);
// this will trigger leader election
@@ -1276,14 +1276,14 @@ public class SimClusterStateProvider implements ClusterStateProvider {
// NOTE: we don't use getProperty because it uses PROPERTY_PROP_PREFIX
Replica leader = s.getLeader();
if (leader == null) {
- LOG.debug("-- no leader in " + s);
+ log.debug("-- no leader in " + s);
continue;
}
cloudManager.getMetricManager().registry(createRegistryName(collection, s.getName(), leader)).counter("UPDATE./update.requests").inc();
ReplicaInfo ri = getReplicaInfo(leader);
Number numDocs = (Number)ri.getVariable("SEARCHER.searcher.numDocs");
if (numDocs == null || numDocs.intValue() <= 0) {
- LOG.debug("-- attempting to delete nonexistent doc " + id + " from " + s.getLeader());
+ log.debug("-- attempting to delete nonexistent doc " + id + " from " + s.getLeader());
continue;
}
modified = true;
@@ -1314,7 +1314,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
for (Slice s : coll.getSlices()) {
Replica leader = s.getLeader();
if (leader == null) {
- LOG.debug("-- no leader in " + s);
+ log.debug("-- no leader in " + s);
continue;
}
@@ -1348,7 +1348,7 @@ public class SimClusterStateProvider implements ClusterStateProvider {
Slice s = router.getTargetSlice(id, null, null, req.getParams(), coll);
Replica leader = s.getLeader();
if (leader == null) {
- LOG.debug("-- no leader in " + s);
+ log.debug("-- no leader in " + s);
continue;
}
cloudManager.getMetricManager().registry(createRegistryName(collection, s.getName(), leader)).counter("UPDATE./update.requests").inc();
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java
index 7a8dbbb..1e99ff2 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistribStateManager.java
@@ -68,7 +68,7 @@ import org.slf4j.LoggerFactory;
* invoked.
*/
public class SimDistribStateManager implements DistribStateManager {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public static final class Node {
ReentrantLock dataLock = new ReentrantLock();
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java
index e9616f0..7168ce9 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimDistributedQueueFactory.java
@@ -51,7 +51,7 @@ import org.slf4j.LoggerFactory;
* exposed anywhere.
*/
public class SimDistributedQueueFactory implements DistributedQueueFactory {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
Map<String, SimDistributedQueue> queues = new ConcurrentHashMap<>();
@@ -190,7 +190,7 @@ public class SimDistributedQueueFactory implements DistributedQueueFactory {
try {
queue.offer(new Pair(String.format(Locale.ROOT, "qn-%010d", seq), data));
seq++;
- LOG.trace("=== offer " + System.nanoTime());
+ log.trace("=== offer " + System.nanoTime());
changed.signalAll();
} finally {
updateLock.unlock();
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java
index 7a346ea..9673fa7 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/SimNodeStateProvider.java
@@ -47,7 +47,7 @@ import org.slf4j.LoggerFactory;
* to setup core-level metrics use {@link SimClusterStateProvider#simSetCollectionValue(String, String, Object, boolean, boolean)}.
*/
public class SimNodeStateProvider implements NodeStateProvider {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private final Map<String, Map<String, Object>> nodeValues = new ConcurrentHashMap<>();
private final SimClusterStateProvider clusterStateProvider;
@@ -164,7 +164,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
* @param node node id
*/
public void simRemoveNodeValues(String node) throws InterruptedException {
- LOG.debug("--removing value for " + node);
+ log.debug("--removing value for " + node);
lock.lockInterruptibly();
try {
Map<String, Object> values = nodeValues.remove(node);
@@ -187,7 +187,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
try {
AtomicBoolean updateRoles = new AtomicBoolean(false);
myNodes.forEach(n -> {
- LOG.debug("- removing dead node values: " + n);
+ log.debug("- removing dead node values: " + n);
Map<String, Object> vals = nodeValues.remove(n);
if (vals.containsKey("nodeRole")) {
updateRoles.set(true);
@@ -253,7 +253,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
for (String tag : tags) {
String[] parts = tag.split(":");
if (parts.length < 3 || !parts[0].equals("metrics")) {
- LOG.warn("Invalid metrics: tag: " + tag);
+ log.warn("Invalid metrics: tag: " + tag);
continue;
}
if (!parts[1].startsWith("solr.core.")) {
@@ -263,7 +263,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
Matcher m = REGISTRY_PATTERN.matcher(parts[1]);
if (!m.matches()) {
- LOG.warn("Invalid registry name: " + parts[1]);
+ log.warn("Invalid registry name: " + parts[1]);
continue;
}
String collection = m.group(1);
@@ -291,7 +291,7 @@ public class SimNodeStateProvider implements NodeStateProvider {
@Override
public Map<String, Object> getNodeValues(String node, Collection<String> tags) {
- LOG.trace("-- requested values for " + node + ": " + tags);
+ log.trace("-- requested values for " + node + ": " + tags);
if (!liveNodesSet.contains(node)) {
throw new RuntimeException("non-live node " + node);
}
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java
index e395985..2cdc456 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestClusterStateProvider.java
@@ -54,7 +54,7 @@ import org.slf4j.LoggerFactory;
* This test compares the cluster state of a real cluster and a simulated one.
*/
public class TestClusterStateProvider extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private static int NODE_COUNT = 3;
private static boolean simulated;
@@ -72,7 +72,7 @@ public class TestClusterStateProvider extends SolrCloudTestCase {
@BeforeClass
public static void setupCluster() throws Exception {
simulated = random().nextBoolean();
- LOG.info("####### Using simulated components? " + simulated);
+ log.info("####### Using simulated components? " + simulated);
configureCluster(NODE_COUNT)
.addConfig("conf", configset("cloud-minimal"))
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java
index 14f7416..74d9bb1 100644
--- a/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java
+++ b/solr/core/src/test/org/apache/solr/cloud/autoscaling/sim/TestDistribStateManager.java
@@ -47,7 +47,7 @@ import org.slf4j.LoggerFactory;
* This test compares a ZK-based {@link DistribStateManager} to the simulated one.
*/
public class TestDistribStateManager extends SolrTestCaseJ4 {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private DistribStateManager stateManager;
private ZkTestServer zkTestServer;
@@ -80,7 +80,7 @@ public class TestDistribStateManager extends SolrTestCaseJ4 {
solrZkClient = new SolrZkClient(zkTestServer.getZkHost(), 30000);
stateManager = new ZkDistribStateManager(solrZkClient);
}
- LOG.info("Using " + stateManager.getClass().getName());
+ log.info("Using " + stateManager.getClass().getName());
}
private DistribStateManager createDistribStateManager() {
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java b/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java
index 652f3f0..362940a 100644
--- a/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java
+++ b/solr/core/src/test/org/apache/solr/search/CurrencyRangeFacetCloudTest.java
@@ -42,7 +42,7 @@ import org.junit.Test;
public class CurrencyRangeFacetCloudTest extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private static final String COLLECTION = MethodHandles.lookup().lookupClass().getName();
private static final String CONF = COLLECTION + "_configSet";
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java
----------------------------------------------------------------------
diff --git a/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java b/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java
index 05c25cf..846977c 100644
--- a/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java
+++ b/solr/core/src/test/org/apache/solr/search/facet/RangeFacetCloudTest.java
@@ -53,7 +53,7 @@ import org.junit.BeforeClass;
*/
public class RangeFacetCloudTest extends SolrCloudTestCase {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private static final String COLLECTION = MethodHandles.lookup().lookupClass().getName();
private static final String CONF = COLLECTION + "_configSet";
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java b/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java
index fa0505e..ccd02eb 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/AutoScalingConfig.java
@@ -43,7 +43,7 @@ import static java.util.stream.Collectors.toList;
* lazily.
*/
public class AutoScalingConfig implements MapWriter {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private final Map<String, Object> jsonMap;
private final boolean empty;
@@ -81,7 +81,7 @@ public class AutoScalingConfig implements MapWriter {
TriggerEventProcessorStage stage = TriggerEventProcessorStage.valueOf(String.valueOf(stageName).toUpperCase(Locale.ROOT));
stages.add(stage);
} catch (Exception e) {
- LOG.warn("Invalid stage name '" + name + "' in listener config, skipping: " + properties);
+ log.warn("Invalid stage name '" + name + "' in listener config, skipping: " + properties);
}
}
listenerClass = (String)this.properties.get(AutoScalingParams.CLASS);
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java b/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java
index 525728e..210e324 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/cloud/autoscaling/Policy.java
@@ -69,7 +69,7 @@ import static org.apache.solr.client.solrj.cloud.autoscaling.Variable.Type.WITH_
*
*/
public class Policy implements MapWriter {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public static final String POLICY = "policy";
public static final String EACH = "#EACH";
@@ -281,7 +281,7 @@ public class Policy implements MapWriter {
return p.compare(r1, r2, false);
});
} catch (Exception e) {
- LOG.error("Exception! prefs = {}, recent r1 = {}, r2 = {}, matrix = {}",
+ log.error("Exception! prefs = {}, recent r1 = {}, r2 = {}, matrix = {}",
clusterPreferences,
lastComparison[0],
lastComparison[1],
@@ -498,9 +498,9 @@ public class Policy implements MapWriter {
this.nodeStateProvider = cloudManager.getNodeStateProvider();
try {
state = cloudManager.getClusterStateProvider().getClusterState();
- LOG.trace("-- session created with cluster state: {}", state);
+ log.trace("-- session created with cluster state: {}", state);
} catch (Exception e) {
- LOG.trace("-- session created, can't obtain cluster state", e);
+ log.trace("-- session created, can't obtain cluster state", e);
}
this.znodeVersion = state != null ? state.getZNodeVersion() : -1;
this.nodes = new ArrayList<>(cloudManager.getClusterStateProvider().getLiveNodes());
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java
index e08f85f..d538b18 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/HttpClientUtil.java
@@ -73,7 +73,7 @@ import org.slf4j.LoggerFactory;
*/
public class HttpClientUtil {
- private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public static final int DEFAULT_CONNECT_TIMEOUT = 60000;
public static final int DEFAULT_SO_TIMEOUT = 600000;
@@ -147,7 +147,7 @@ public class HttpClientUtil {
// Configure the HttpClientBuilder if user has specified the factory type.
String factoryClassName = System.getProperty(SYS_PROP_HTTP_CLIENT_BUILDER_FACTORY);
if (factoryClassName != null) {
- logger.debug ("Using " + factoryClassName);
+ log.debug ("Using " + factoryClassName);
try {
HttpClientBuilderFactory factory = (HttpClientBuilderFactory)Class.forName(factoryClassName).newInstance();
httpClientBuilder = factory.getHttpClientBuilder(Optional.of(SolrHttpClientBuilder.create()));
@@ -176,7 +176,7 @@ public class HttpClientUtil {
try {
interceptor.process(request, context);
} catch (Exception e) {
- logger.error("", e);
+ log.error("", e);
}
}
});
@@ -234,7 +234,7 @@ public class HttpClientUtil {
} else {
sslConnectionSocketFactory = new SSLConnectionSocketFactory(SSLContexts.createSystemDefault(),
NoopHostnameVerifier.INSTANCE);
- logger.debug(HttpClientUtil.SYS_PROP_CHECK_PEER_NAME + "is false, hostname checks disabled.");
+ log.debug(HttpClientUtil.SYS_PROP_CHECK_PEER_NAME + "is false, hostname checks disabled.");
}
builder.register("https", sslConnectionSocketFactory);
@@ -268,8 +268,8 @@ public class HttpClientUtil {
public static CloseableHttpClient createClient(final SolrParams params, PoolingHttpClientConnectionManager cm, boolean sharedConnectionManager, HttpRequestExecutor httpRequestExecutor) {
final ModifiableSolrParams config = new ModifiableSolrParams(params);
- if (logger.isDebugEnabled()) {
- logger.debug("Creating new http client, config:" + config);
+ if (log.isDebugEnabled()) {
+ log.debug("Creating new http client, config:" + config);
}
cm.setMaxTotal(params.getInt(HttpClientUtil.PROP_MAX_CONNECTIONS, 10000));
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java
index 1bcf96b..afa2ef6 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/impl/Krb5HttpClientBuilder.java
@@ -50,7 +50,7 @@ import org.slf4j.LoggerFactory;
public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
public static final String LOGIN_CONFIG_PROP = "java.security.auth.login.config";
- private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private static Configuration jaasConfig = new SolrJaasConfiguration();
@@ -85,7 +85,7 @@ public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
String configValue = System.getProperty(LOGIN_CONFIG_PROP);
if (configValue != null) {
- logger.info("Setting up SPNego auth with config: " + configValue);
+ log.info("Setting up SPNego auth with config: " + configValue);
final String useSubjectCredsProp = "javax.security.auth.useSubjectCredsOnly";
String useSubjectCredsVal = System.getProperty(useSubjectCredsProp);
@@ -97,7 +97,7 @@ public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
else if (!useSubjectCredsVal.toLowerCase(Locale.ROOT).equals("false")) {
// Don't overwrite the prop value if it's already been written to something else,
// but log because it is likely the Credentials won't be loaded correctly.
- logger.warn("System Property: " + useSubjectCredsProp + " set to: " + useSubjectCredsVal
+ log.warn("System Property: " + useSubjectCredsProp + " set to: " + useSubjectCredsVal
+ " not false. SPNego authentication may not be successful.");
}
@@ -139,7 +139,7 @@ public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
HttpClientUtil.addRequestInterceptor(bufferedEntityInterceptor);
}
} else {
- logger.warn("{} is configured without specifying system property '{}'",
+ log.warn("{} is configured without specifying system property '{}'",
getClass().getName(), LOGIN_CONFIG_PROP);
}
@@ -176,11 +176,11 @@ public class Krb5HttpClientBuilder implements HttpClientBuilderFactory {
public AppConfigurationEntry[] getAppConfigurationEntry(String appName) {
if (baseConfig == null) return null;
- logger.debug("Login prop: "+System.getProperty(LOGIN_CONFIG_PROP));
+ log.debug("Login prop: "+System.getProperty(LOGIN_CONFIG_PROP));
String clientAppName = System.getProperty("solr.kerberos.jaas.appname", "Client");
if (initiateAppNames.contains(appName)) {
- logger.debug("Using AppConfigurationEntry for appName '"+clientAppName+"' instead of: " + appName);
+ log.debug("Using AppConfigurationEntry for appName '"+clientAppName+"' instead of: " + appName);
return baseConfig.getAppConfigurationEntry(clientAppName);
}
return baseConfig.getAppConfigurationEntry(appName);
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java
index dc229bf..b29ea09 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/CommitStream.java
@@ -43,7 +43,7 @@ import org.slf4j.LoggerFactory;
*/
public class CommitStream extends TupleStream implements Expressible {
private static final long serialVersionUID = 1L;
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
// Part of expression / passed in
private String collection;
@@ -252,7 +252,7 @@ public class CommitStream extends TupleStream implements Expressible {
try {
clientCache.getCloudSolrClient(zkHost).commit(collection, waitFlush, waitSearcher, softCommit);
} catch (SolrServerException | IOException e) {
- LOG.warn(String.format(Locale.ROOT, "Unable to commit documents to collection '%s' due to unexpected error.", collection), e);
+ log.warn(String.format(Locale.ROOT, "Unable to commit documents to collection '%s' due to unexpected error.", collection), e);
String className = e.getClass().getName();
String message = e.getMessage();
throw new IOException(String.format(Locale.ROOT,"Unexpected error when committing documents to collection %s- %s:%s", collection, className, message));
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java
index a4f528b..9d02ec2 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/DaemonStream.java
@@ -60,7 +60,7 @@ public class DaemonStream extends TupleStream implements Expressible {
private Map<String, DaemonStream> daemons;
private boolean terminate;
private boolean closed = false;
- private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public DaemonStream(StreamExpression expression, StreamFactory factory) throws IOException{
@@ -329,14 +329,14 @@ public class DaemonStream extends TupleStream implements Expressible {
}
} catch (IOException e) {
exception = e;
- logger.error("Error in DaemonStream:" + id, e);
+ log.error("Error in DaemonStream:" + id, e);
++errors;
if (errors > 100) {
- logger.error("Too many consectutive errors. Stopping DaemonStream:" + id);
+ log.error("Too many consectutive errors. Stopping DaemonStream:" + id);
break OUTER;
}
} catch (Throwable t) {
- logger.error("Fatal Error in DaemonStream:" + id, t);
+ log.error("Fatal Error in DaemonStream:" + id, t);
//For anything other then IOException break out of the loop and shutdown the thread.
break OUTER;
} finally {
@@ -345,7 +345,7 @@ public class DaemonStream extends TupleStream implements Expressible {
} catch (IOException e1) {
if (exception == null) {
exception = e1;
- logger.error("Error in DaemonStream:" + id, e1);
+ log.error("Error in DaemonStream:" + id, e1);
break OUTER;
}
}
@@ -357,7 +357,7 @@ public class DaemonStream extends TupleStream implements Expressible {
try {
Thread.sleep(sleepMillis);
} catch (InterruptedException e) {
- logger.error("Error in DaemonStream:" + id, e);
+ log.error("Error in DaemonStream:" + id, e);
break OUTER;
}
}
@@ -370,7 +370,7 @@ public class DaemonStream extends TupleStream implements Expressible {
try {
queue.put(tuple);
} catch (InterruptedException e) {
- logger.error("Error in DaemonStream:"+id, e);
+ log.error("Error in DaemonStream:"+id, e);
}
}
setStopTime(new Date().getTime());
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java
index 2a1b491..bee208c 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/ExecutorStream.java
@@ -55,7 +55,7 @@ import static org.apache.solr.common.params.CommonParams.ID;
public class ExecutorStream extends TupleStream implements Expressible {
- private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
private TupleStream stream;
@@ -148,7 +148,7 @@ public class ExecutorStream extends TupleStream implements Expressible {
try {
executorService.awaitTermination(Long.MAX_VALUE, TimeUnit.SECONDS);
} catch(InterruptedException e) {
- logger.error("Interrupted while waiting for termination", e);
+ log.error("Interrupted while waiting for termination", e);
}
}
@@ -214,12 +214,12 @@ public class ExecutorStream extends TupleStream implements Expressible {
}
}
} catch (Exception e) {
- logger.error("Executor Error: id="+id+" expr_s="+expr, e);
+ log.error("Executor Error: id="+id+" expr_s="+expr, e);
} finally {
try {
stream.close();
} catch (Exception e1) {
- logger.error("Executor Error", e1);
+ log.error("Executor Error", e1);
}
}
}
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java
index a08f838..c00de10 100644
--- a/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java
+++ b/solr/solrj/src/java/org/apache/solr/client/solrj/io/stream/UpdateStream.java
@@ -50,7 +50,7 @@ import static org.apache.solr.common.params.CommonParams.VERSION_FIELD;
* @since 6.0.0
*/
public class UpdateStream extends TupleStream implements Expressible {
- private static final Logger LOG = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
public static String BATCH_INDEXED_FIELD_NAME = "batchIndexed"; // field name in summary tuple for #docs updated in batch
private String collection;
@@ -281,7 +281,7 @@ public class UpdateStream extends TupleStream implements Expressible {
}
}
}
- LOG.debug("Tuple [{}] was converted into SolrInputDocument [{}].", tuple, doc);
+ log.debug("Tuple [{}] was converted into SolrInputDocument [{}].", tuple, doc);
return doc;
}
@@ -300,7 +300,7 @@ public class UpdateStream extends TupleStream implements Expressible {
try {
cloudSolrClient.add(collection, documentBatch);
} catch (SolrServerException | IOException e) {
- LOG.warn("Unable to add documents to collection due to unexpected error.", e);
+ log.warn("Unable to add documents to collection due to unexpected error.", e);
String className = e.getClass().getName();
String message = e.getMessage();
throw new IOException(String.format(Locale.ROOT,"Unexpected error when adding documents to collection %s- %s:%s", collection, className, message));
http://git-wip-us.apache.org/repos/asf/lucene-solr/blob/8cde1277/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java
----------------------------------------------------------------------
diff --git a/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java b/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java
index 015793c..bf4a1ed 100644
--- a/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java
+++ b/solr/solrj/src/java/org/apache/solr/common/cloud/ZkConfigManager.java
@@ -36,7 +36,7 @@ import java.util.regex.Pattern;
*/
public class ZkConfigManager {
- private static final Logger logger = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
+ private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass());
/** ZkNode where named configs are stored */
public static final String CONFIGS_ZKNODE = "/configs";
@@ -139,7 +139,7 @@ public class ZkConfigManager {
List<String> children = zkClient.getChildren(fromZkPath + "/" + file, null, true);
if (children.size() == 0) {
final String toZkFilePath = toZkPath + "/" + file;
- logger.info("Copying zk node {} to {}",
+ log.info("Copying zk node {} to {}",
fromZkPath + "/" + file, toZkFilePath);
byte[] data = zkClient.getData(fromZkPath + "/" + file, null, null, true);
zkClient.makePath(toZkFilePath, data, true);