You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2018/06/22 15:58:16 UTC

[JENKINS] Lucene-Solr-master-Linux (32bit/jdk1.8.0_172) - Build # 22298 - Unstable!

Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22298/
Java: 32bit/jdk1.8.0_172 -server -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  org.apache.solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest.test

Error Message:
Timed out waiting for replica core_node52 (1529681195850) to replicate from leader core_node44 (0)

Stack Trace:
java.lang.AssertionError: Timed out waiting for replica core_node52 (1529681195850) to replicate from leader core_node44 (0)
	at __randomizedtesting.SeedInfo.seed([809835CCA1EF3CC3:8CC0A160F13513B]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForReplicationFromReplicas(AbstractFullDistribZkTestBase.java:2146)
	at org.apache.solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest.test(ChaosMonkeySafeLeaderWithPullReplicasTest.java:211)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:993)
	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:968)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 14375 lines...]
   [junit4] Suite: org.apache.solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest
   [junit4]   2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/init-core-data-001
   [junit4]   2> 1524922 INFO  (SUITE-ChaosMonkeySafeLeaderWithPullReplicasTest-seed#[809835CCA1EF3CC3]-worker) [    ] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /y/b
   [junit4]   2> 1524923 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.ChaosMonkeySafeLeaderWithPullReplicasTest Starting ChaosMonkey test with 2 shards and 6 nodes
   [junit4]   2> 1524924 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 1524924 INFO  (Thread-2837) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 1524924 INFO  (Thread-2837) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 1524926 ERROR (Thread-2837) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 1525028 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.ZkTestServer start zk server on port:42257
   [junit4]   2> 1525048 INFO  (zkConnectionManagerCallback-4028-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1525058 WARN  (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [    ] o.a.z.s.NIOServerCnxn Unable to read additional data from client sessionid 0x100aa9a5de80000, likely client has closed socket
   [junit4]   2> 1525066 INFO  (zkConnectionManagerCallback-4030-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1525069 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml
   [junit4]   2> 1525069 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/schema15.xml to /configs/conf1/schema.xml
   [junit4]   2> 1525070 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 1525071 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
   [junit4]   2> 1525071 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
   [junit4]   2> 1525072 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
   [junit4]   2> 1525072 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml
   [junit4]   2> 1525073 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 1525073 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 1525074 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
   [junit4]   2> 1525074 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractZkTestCase put /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
   [junit4]   2> 1525085 INFO  (zkConnectionManagerCallback-4034-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1525086 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly asked otherwise
   [junit4]   2> 1525193 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
   [junit4]   2> 1525194 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1525194 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1525194 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1525196 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@106009b{/y/b,null,AVAILABLE}
   [junit4]   2> 1525201 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.AbstractConnector Started ServerConnector@a9345{SSL,[ssl, http/1.1]}{127.0.0.1:40445}
   [junit4]   2> 1525201 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.Server Started @1525237ms
   [junit4]   2> 1525201 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/tempDir-001/control/data, hostContext=/y/b, hostPort=40445, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/../../../../../../../../../home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/control-001/cores}
   [junit4]   2> 1525201 ERROR (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1525201 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1525201 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 8.0.0
   [junit4]   2> 1525201 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1525202 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1525202 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-22T15:25:31.897Z
   [junit4]   2> 1525216 INFO  (zkConnectionManagerCallback-4036-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1525217 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 1525217 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/control-001/solr.xml
   [junit4]   2> 1525221 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1525221 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1525222 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1526305 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:42257/solr
   [junit4]   2> 1526306 INFO  (zkConnectionManagerCallback-4040-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1526307 INFO  (zkConnectionManagerCallback-4042-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1526343 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1526344 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:40445_y%2Fb
   [junit4]   2> 1526344 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.c.Overseer Overseer (id=72245174015098885-127.0.0.1:40445_y%2Fb-n_0000000000) starting
   [junit4]   2> 1526349 INFO  (zkConnectionManagerCallback-4049-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1526349 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:42257/solr ready
   [junit4]   2> 1526350 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:40445_y%2Fb
   [junit4]   2> 1526351 INFO  (zkCallback-4048-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1526351 INFO  (zkCallback-4041-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1526364 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1526377 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1526383 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1526383 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1526384 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/../../../../../../../../../home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/control-001/cores
   [junit4]   2> 1526402 INFO  (zkConnectionManagerCallback-4055-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1526402 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1526403 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:42257/solr ready
   [junit4]   2> 1526441 INFO  (qtp11920436-14736) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:40445_y%252Fb&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1526444 INFO  (OverseerThreadFactory-6495-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection control_collection
   [junit4]   2> 1526550 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb    x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1526550 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb    x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1527560 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1527569 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test
   [junit4]   2> 1527630 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 1527638 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' using configuration from collection control_collection, trusted=true
   [junit4]   2> 1527638 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.control_collection.shard1.replica_n1' (registry 'solr.core.control_collection.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1527638 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 1527638 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/control-001/cores/control_collection_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/../../../../../../../../../home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/control-001/cores/control_collection_shard1_replica_n1/data/]
   [junit4]   2> 1527642 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: minMergeSize=1000, mergeFactor=26, maxMergeSize=9223372036854775807, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0]
   [junit4]   2> 1527662 WARN  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
   [junit4]   2> 1527693 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 1527693 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 1527694 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: if uncommited for 15000ms; 
   [junit4]   2> 1527694 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 1527695 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=36, maxMergedSegmentMB=50.109375, floorSegmentMB=1.4150390625, forceMergeDeletesPctAllowed=0.5326033755995785, segmentsPerTier=13.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, reclaimDeletesWeight=1.6538614152617783
   [junit4]   2> 1527695 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@8d11cc[control_collection_shard1_replica_n1] main]
   [junit4]   2> 1527696 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 1527696 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 1527696 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 1527697 INFO  (searcherExecutor-6500-thread-1-processing-n:127.0.0.1:40445_y%2Fb x:control_collection_shard1_replica_n1 c:control_collection s:shard1) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [control_collection_shard1_replica_n1] Registered new searcher Searcher@8d11cc[control_collection_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 1527697 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603986925176225792
   [junit4]   2> 1527701 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, version=0}
   [junit4]   2> 1527702 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 1527702 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 1527702 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:40445/y/b/control_collection_shard1_replica_n1/
   [junit4]   2> 1527702 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 1527702 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:40445/y/b/control_collection_shard1_replica_n1/ has no replicas
   [junit4]   2> 1527702 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 1527703 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:40445/y/b/control_collection_shard1_replica_n1/ shard1
   [junit4]   2> 1527804 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 1527805 INFO  (qtp11920436-14732) [n:127.0.0.1:40445_y%2Fb c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1256
   [junit4]   2> 1527807 INFO  (qtp11920436-14736) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 1527906 INFO  (zkCallback-4041-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 1528445 INFO  (OverseerCollectionConfigSetProcessor-72245174015098885-127.0.0.1:40445_y%2Fb-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 1528807 INFO  (qtp11920436-14736) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:40445_y%252Fb&wt=javabin&version=2} status=0 QTime=2365
   [junit4]   2> 1528810 INFO  (zkConnectionManagerCallback-4060-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1528810 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1528811 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:42257/solr ready
   [junit4]   2> 1528811 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection loss:false
   [junit4]   2> 1528812 INFO  (qtp11920436-14737) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=2&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1528813 INFO  (OverseerThreadFactory-6495-thread-2) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection collection1
   [junit4]   2> 1528814 WARN  (OverseerThreadFactory-6495-thread-2) [    ] o.a.s.c.a.c.CreateCollectionCmd It is unusual to create a collection (collection1) without cores.
   [junit4]   2> 1529017 INFO  (qtp11920436-14737) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 1529018 INFO  (qtp11920436-14737) [n:127.0.0.1:40445_y%2Fb    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=2&wt=javabin&version=2} status=0 QTime=205
   [junit4]   2> 1529092 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-1-001 of type NRT
   [junit4]   2> 1529093 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
   [junit4]   2> 1529094 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1529094 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1529094 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1529094 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@7aff33{/y/b,null,AVAILABLE}
   [junit4]   2> 1529095 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.AbstractConnector Started ServerConnector@118bf3{SSL,[ssl, http/1.1]}{127.0.0.1:34197}
   [junit4]   2> 1529095 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.Server Started @1529131ms
   [junit4]   2> 1529095 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/tempDir-001/jetty1, replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/y/b, hostPort=34197, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-1-001/cores}
   [junit4]   2> 1529095 ERROR (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1529095 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1529095 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 8.0.0
   [junit4]   2> 1529095 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1529095 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1529095 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-22T15:25:35.790Z
   [junit4]   2> 1529096 INFO  (zkConnectionManagerCallback-4062-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1529097 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 1529097 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-1-001/solr.xml
   [junit4]   2> 1529099 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1529099 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1529100 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1530104 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:42257/solr
   [junit4]   2> 1530105 INFO  (zkConnectionManagerCallback-4066-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1530106 INFO  (zkConnectionManagerCallback-4068-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1530109 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1530110 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1530111 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1530111 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34197_y%2Fb
   [junit4]   2> 1530111 INFO  (zkCallback-4048-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1530111 INFO  (zkCallback-4041-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1530111 INFO  (zkCallback-4059-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1530111 INFO  (zkCallback-4067-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1530125 INFO  (zkConnectionManagerCallback-4075-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1530127 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 1530127 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:42257/solr ready
   [junit4]   2> 1530127 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1530142 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1530150 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1530150 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1530152 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-1-001/cores
   [junit4]   2> 1530204 INFO  (qtp20882808-14794) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:34197_y%252Fb&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1530206 INFO  (OverseerCollectionConfigSetProcessor-72245174015098885-127.0.0.1:40445_y%2Fb-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000002 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 1530206 INFO  (OverseerThreadFactory-6495-thread-3) [ c:collection1 s:shard2  ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:34197_y%2Fb for creating new replica
   [junit4]   2> 1530210 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb    x:collection1_shard2_replica_n41] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n41&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1531256 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1531264 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema [collection1_shard2_replica_n41] Schema name=test
   [junit4]   2> 1531313 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 1531320 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard2_replica_n41' using configuration from collection collection1, trusted=true
   [junit4]   2> 1531321 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard2.replica_n41' (registry 'solr.core.collection1.shard2.replica_n41') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1531321 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 1531321 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SolrCore [[collection1_shard2_replica_n41] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-1-001/cores/collection1_shard2_replica_n41], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-1-001/cores/collection1_shard2_replica_n41/data/]
   [junit4]   2> 1531324 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: minMergeSize=1000, mergeFactor=26, maxMergeSize=9223372036854775807, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0]
   [junit4]   2> 1531339 WARN  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
   [junit4]   2> 1531377 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 1531377 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 1531378 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.CommitTracker Hard AutoCommit: if uncommited for 15000ms; 
   [junit4]   2> 1531378 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 1531380 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=36, maxMergedSegmentMB=50.109375, floorSegmentMB=1.4150390625, forceMergeDeletesPctAllowed=0.5326033755995785, segmentsPerTier=13.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, reclaimDeletesWeight=1.6538614152617783
   [junit4]   2> 1531380 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.s.SolrIndexSearcher Opening [Searcher@6cbbd9[collection1_shard2_replica_n41] main]
   [junit4]   2> 1531381 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 1531382 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 1531382 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 1531382 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603986929040228352
   [junit4]   2> 1531383 INFO  (searcherExecutor-6514-thread-1-processing-n:127.0.0.1:34197_y%2Fb x:collection1_shard2_replica_n41 c:collection1 s:shard2) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SolrCore [collection1_shard2_replica_n41] Registered new searcher Searcher@6cbbd9[collection1_shard2_replica_n41] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 1531389 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard2 to Terms{values={core_node42=0}, version=0}
   [junit4]   2> 1531390 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 1531390 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 1531390 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:34197/y/b/collection1_shard2_replica_n41/
   [junit4]   2> 1531390 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 1531390 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy https://127.0.0.1:34197/y/b/collection1_shard2_replica_n41/ has no replicas
   [junit4]   2> 1531390 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 1531392 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:34197/y/b/collection1_shard2_replica_n41/ shard2
   [junit4]   2> 1531493 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 1531494 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n41&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1285
   [junit4]   2> 1531496 INFO  (qtp20882808-14794) [n:127.0.0.1:34197_y%2Fb c:collection1   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:34197_y%252Fb&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2} status=0 QTime=1291
   [junit4]   2> 1531570 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-2-001 of type NRT
   [junit4]   2> 1531571 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
   [junit4]   2> 1531572 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1531572 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1531572 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1531572 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@1dcdd3e{/y/b,null,AVAILABLE}
   [junit4]   2> 1531572 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.AbstractConnector Started ServerConnector@1a5092{SSL,[ssl, http/1.1]}{127.0.0.1:37011}
   [junit4]   2> 1531572 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.Server Started @1531608ms
   [junit4]   2> 1531572 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/tempDir-001/jetty2, solrconfig=solrconfig.xml, hostContext=/y/b, hostPort=37011, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-2-001/cores}
   [junit4]   2> 1531573 ERROR (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1531573 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1531573 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 8.0.0
   [junit4]   2> 1531573 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1531573 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1531573 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-22T15:25:38.268Z
   [junit4]   2> 1531574 INFO  (zkConnectionManagerCallback-4078-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1531575 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 1531575 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-2-001/solr.xml
   [junit4]   2> 1531577 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1531577 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1531577 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1531595 INFO  (zkCallback-4067-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 1532207 INFO  (OverseerCollectionConfigSetProcessor-72245174015098885-127.0.0.1:40445_y%2Fb-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000004 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 1532725 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:42257/solr
   [junit4]   2> 1532725 INFO  (zkConnectionManagerCallback-4082-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1532727 INFO  (zkConnectionManagerCallback-4084-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1532731 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 1532732 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1532732 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1532733 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:37011_y%2Fb
   [junit4]   2> 1532733 INFO  (zkCallback-4041-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1532733 INFO  (zkCallback-4067-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1532733 INFO  (zkCallback-4048-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1532733 INFO  (zkCallback-4074-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1532733 INFO  (zkCallback-4059-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1532733 INFO  (zkCallback-4083-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1532746 INFO  (zkConnectionManagerCallback-4091-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1532747 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 1532748 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:42257/solr ready
   [junit4]   2> 1532748 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1532763 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1532770 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1532770 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1532771 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:37011_y%2Fb    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-2-001/cores
   [junit4]   2> 1532796 INFO  (qtp20882808-14795) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:37011_y%252Fb&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1532798 INFO  (OverseerThreadFactory-6495-thread-4) [ c:collection1 s:shard1  ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:37011_y%2Fb for creating new replica
   [junit4]   2> 1532831 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb    x:collection1_shard1_replica_n43] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n43&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1532935 INFO  (zkCallback-4067-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [3])
   [junit4]   2> 1533844 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1533852 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.s.IndexSchema [collection1_shard1_replica_n43] Schema name=test
   [junit4]   2> 1533909 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 1533917 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard1_replica_n43' using configuration from collection collection1, trusted=true
   [junit4]   2> 1533917 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard1.replica_n43' (registry 'solr.core.collection1.shard1.replica_n43') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1533918 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 1533918 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.SolrCore [[collection1_shard1_replica_n43] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-2-001/cores/collection1_shard1_replica_n43], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-2-001/cores/collection1_shard1_replica_n43/data/]
   [junit4]   2> 1533921 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: minMergeSize=1000, mergeFactor=26, maxMergeSize=9223372036854775807, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0]
   [junit4]   2> 1533937 WARN  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
   [junit4]   2> 1533973 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 1533973 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 1533974 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.u.CommitTracker Hard AutoCommit: if uncommited for 15000ms; 
   [junit4]   2> 1533974 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 1533975 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=36, maxMergedSegmentMB=50.109375, floorSegmentMB=1.4150390625, forceMergeDeletesPctAllowed=0.5326033755995785, segmentsPerTier=13.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, reclaimDeletesWeight=1.6538614152617783
   [junit4]   2> 1533975 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.s.SolrIndexSearcher Opening [Searcher@1c5aa79[collection1_shard1_replica_n43] main]
   [junit4]   2> 1533976 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 1533976 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 1533977 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 1533977 INFO  (searcherExecutor-6528-thread-1-processing-n:127.0.0.1:37011_y%2Fb x:collection1_shard1_replica_n43 c:collection1 s:shard1) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.SolrCore [collection1_shard1_replica_n43] Registered new searcher Searcher@1c5aa79[collection1_shard1_replica_n43] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 1533977 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603986931761283072
   [junit4]   2> 1533981 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard1 to Terms{values={core_node44=0}, version=0}
   [junit4]   2> 1533982 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 1533982 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 1533982 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:37011/y/b/collection1_shard1_replica_n43/
   [junit4]   2> 1533982 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 1533982 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.SyncStrategy https://127.0.0.1:37011/y/b/collection1_shard1_replica_n43/ has no replicas
   [junit4]   2> 1533982 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 1533983 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:37011/y/b/collection1_shard1_replica_n43/ shard1
   [junit4]   2> 1534084 INFO  (zkCallback-4067-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [3])
   [junit4]   2> 1534084 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 1534085 INFO  (qtp4881791-14834) [n:127.0.0.1:37011_y%2Fb c:collection1 s:shard1  x:collection1_shard1_replica_n43] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard1_replica_n43&action=CREATE&collection=collection1&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1254
   [junit4]   2> 1534086 INFO  (qtp20882808-14795) [n:127.0.0.1:34197_y%2Fb c:collection1   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:37011_y%252Fb&action=ADDREPLICA&collection=collection1&shard=shard1&type=NRT&wt=javabin&version=2} status=0 QTime=1290
   [junit4]   2> 1534163 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 3 in directory /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-3-001 of type NRT
   [junit4]   2> 1534164 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 1.8.0_172-b11
   [junit4]   2> 1534166 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1534166 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1534166 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1534166 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@1841991{/y/b,null,AVAILABLE}
   [junit4]   2> 1534167 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.AbstractConnector Started ServerConnector@170a62{SSL,[ssl, http/1.1]}{127.0.0.1:44237}
   [junit4]   2> 1534167 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.e.j.s.Server Started @1534203ms
   [junit4]   2> 1534167 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/tempDir-001/jetty3, solrconfig=solrconfig.xml, hostContext=/y/b, hostPort=44237, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-3-001/cores}
   [junit4]   2> 1534167 ERROR (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1534167 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1534167 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 8.0.0
   [junit4]   2> 1534167 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1534167 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1534168 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-22T15:25:40.863Z
   [junit4]   2> 1534169 INFO  (zkConnectionManagerCallback-4094-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1534170 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 1534170 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-3-001/solr.xml
   [junit4]   2> 1534172 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1534172 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1534173 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1534185 INFO  (zkCallback-4083-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [3])
   [junit4]   2> 1534185 INFO  (zkCallback-4067-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [3])
   [junit4]   2> 1534581 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:42257/solr
   [junit4]   2> 1534582 INFO  (zkConnectionManagerCallback-4098-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1534583 INFO  (zkConnectionManagerCallback-4100-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1534586 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 1534586 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1534587 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1534587 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:44237_y%2Fb
   [junit4]   2> 1534588 INFO  (zkCallback-4048-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 1534588 INFO  (zkCallback-4041-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 1534588 INFO  (zkCallback-4067-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 1534588 INFO  (zkCallback-4083-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 1534588 INFO  (zkCallback-4059-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 1534588 INFO  (zkCallback-4074-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 1534588 INFO  (zkCallback-4090-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 1534588 INFO  (zkCallback-4099-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 1534601 INFO  (zkConnectionManagerCallback-4107-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1534602 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
   [junit4]   2> 1534602 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:42257/solr ready
   [junit4]   2> 1534603 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1534616 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1534624 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1534624 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1534625 INFO  (TEST-ChaosMonkeySafeLeaderWithPullReplicasTest.test-seed#[809835CCA1EF3CC3]) [n:127.0.0.1:44237_y%2Fb    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-3-001/cores
   [junit4]   2> 1534655 INFO  (qtp20882808-14794) [n:127.0.0.1:34197_y%2Fb    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:44237_y%252Fb&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1534657 INFO  (OverseerCollectionConfigSetProcessor-72245174015098885-127.0.0.1:40445_y%2Fb-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000006 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 1534657 INFO  (OverseerThreadFactory-6495-thread-5) [ c:collection1 s:shard2  ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:44237_y%2Fb for creating new replica
   [junit4]   2> 1534691 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb    x:collection1_shard2_replica_n45] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n45&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1534794 INFO  (zkCallback-4067-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [4])
   [junit4]   2> 1534794 INFO  (zkCallback-4083-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/collection1/state.json] for collection [collection1] has occurred - updating... (live nodes size: [4])
   [junit4]   2> 1535705 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1535714 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.s.IndexSchema [collection1_shard2_replica_n45] Schema name=test
   [junit4]   2> 1535761 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 1535769 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard2_replica_n45' using configuration from collection collection1, trusted=true
   [junit4]   2> 1535770 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard2.replica_n45' (registry 'solr.core.collection1.shard2.replica_n45') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@144c3ed
   [junit4]   2> 1535770 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 1535770 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.c.SolrCore [[collection1_shard2_replica_n45] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-3-001/cores/collection1_shard2_replica_n45], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J1/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001/shard-3-001/cores/collection1_shard2_replica_n45/data/]
   [junit4]   2> 1535773 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.LogDocMergePolicy: [LogDocMergePolicy: minMergeSize=1000, mergeFactor=26, maxMergeSize=9223372036854775807, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=false, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0]
   [junit4]   2> 1535791 WARN  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
   [junit4]   2> 1535823 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 1535823 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 1535824 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.u.CommitTracker Hard AutoCommit: if uncommited for 15000ms; 
   [junit4]   2> 1535824 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 1535826 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=36, maxMergedSegmentMB=50.109375, floorSegmentMB=1.4150390625, forceMergeDeletesPctAllowed=0.5326033755995785, segmentsPerTier=13.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, reclaimDeletesWeight=1.6538614152617783
   [junit4]   2> 1535826 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.s.SolrIndexSearcher Opening [Searcher@16eccd1[collection1_shard2_replica_n45] main]
   [junit4]   2> 1535827 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 1535827 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 1535827 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 1535828 INFO  (searcherExecutor-6542-thread-1-processing-n:127.0.0.1:44237_y%2Fb x:collection1_shard2_replica_n45 c:collection1 s:shard2) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.c.SolrCore [collection1_shard2_replica_n45] Registered new searcher Searcher@16eccd1[collection1_shard2_replica_n45] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 1535828 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1603986933702197248
   [junit4]   2> 1535831 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard2 to Terms{values={core_node42=0, core_node46=0}, version=1}
   [junit4]   2> 1535831 INFO  (qtp20619057-14872) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2  x:collection1_shard2_replica_n45] o.a.s.c.ZkController Core needs to recover:collection1_shard2_replica_n45
   [junit4]   2> 1535831 INFO  (updateExecutor-4095-thread-1-processing-n:127.0.0.1:44237_y%2Fb x:collection1_shard2_replica_n45 c:collection1 s:shard2) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.u.DefaultSolrCoreState Running recovery
   [junit4]   2> 1535832 INFO  (recoveryExecutor-4096-thread-1-processing-n:127.0.0.1:44237_y%2Fb x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.RecoveryStrategy Starting recovery process. recoveringAfterStartup=true
   [junit4]   2> 1535832 INFO  (recoveryExecutor-4096-thread-1-processing-n:127.0.0.1:44237_y%2Fb x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.RecoveryStrategy ###### startupVersions=[[]]
   [junit4]   2> 1535835 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2 r:core_node42 x:collection1_shard2_replica_n41] o.a.s.c.S.Request [collection1_shard2_replica_n41]  webapp=/y/b path=/admin/ping params={wt=javabin&version=2} hits=0 status=0 QTime=0
   [junit4]   2> 1535835 INFO  (qtp20882808-14823) [n:127.0.0.1:34197_y%2Fb c:collection1 s:shard2 r:core_node42 x:collection1_shard2_replica_n41] o.a.s.c.S.Request [collection1_shard2_replica_n41]  webapp=/y/b path=/admin/ping params={wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 1535836 INFO  (recoveryExecutor-4096-thread-1-processing-n:127.0.0.1:44237_y%2Fb x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.c.RecoveryStrategy Begin buffering updates. core=[collection1_shard2_replica_n45]
   [junit4]   2> 1535836 INFO  (recoveryExecutor-4096-thread-1-processing-n:127.0.0.1:44237_y%2Fb x:collection1_shard2_replica_n45 c:collection1 s:shard2 r:core_node46) [n:127.0.0.1:44237_y%2Fb c:collection1 s:shard2 r:core_node46 x:collection1_shard2_replica_n45] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 1535836 INFO  (recoveryExecutor-4096-thread-1-processing-n:127.0.0.1:44237_y%2F

[...truncated too long message...]

ion1//collections/collection1/state.json/37)={
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "shards":{
   [junit4]   2>     "shard1":{
   [junit4]   2>       "range":"80000000-ffffffff",
   [junit4]   2>       "state":"active",
   [junit4]   2>       "replicas":{
   [junit4]   2>         "core_node44":{
   [junit4]   2>           "core":"collection1_shard1_replica_n43",
   [junit4]   2>           "base_url":"https://127.0.0.1:45485/y/b",
   [junit4]   2>           "node_name":"127.0.0.1:45485_y%2Fb",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT"},
   [junit4]   2>         "core_node48":{
   [junit4]   2>           "core":"collection1_shard1_replica_n47",
   [junit4]   2>           "base_url":"https://127.0.0.1:45131/y/b",
   [junit4]   2>           "node_name":"127.0.0.1:45131_y%2Fb",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT"},
   [junit4]   2>         "core_node52":{
   [junit4]   2>           "core":"collection1_shard1_replica_p51",
   [junit4]   2>           "base_url":"https://127.0.0.1:33589/y/b",
   [junit4]   2>           "node_name":"127.0.0.1:33589_y%2Fb",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"PULL"}}},
   [junit4]   2>     "shard2":{
   [junit4]   2>       "range":"0-7fffffff",
   [junit4]   2>       "state":"active",
   [junit4]   2>       "replicas":{
   [junit4]   2>         "core_node42":{
   [junit4]   2>           "core":"collection1_shard2_replica_n41",
   [junit4]   2>           "base_url":"https://127.0.0.1:36147/y/b",
   [junit4]   2>           "node_name":"127.0.0.1:36147_y%2Fb",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT"},
   [junit4]   2>         "core_node46":{
   [junit4]   2>           "core":"collection1_shard2_replica_n45",
   [junit4]   2>           "base_url":"https://127.0.0.1:42397/y/b",
   [junit4]   2>           "node_name":"127.0.0.1:42397_y%2Fb",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT"},
   [junit4]   2>         "core_node50":{
   [junit4]   2>           "core":"collection1_shard2_replica_p49",
   [junit4]   2>           "base_url":"https://127.0.0.1:46771/y/b",
   [junit4]   2>           "node_name":"127.0.0.1:46771_y%2Fb",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"PULL"}}}},
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "tlogReplicas":"0"} with live_nodes=[]
   [junit4]   2> 	at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:861)
   [junit4]   2> 	at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:845)
   [junit4]   2> 	at org.apache.solr.handler.IndexFetcher.getLeaderReplica(IndexFetcher.java:689)
   [junit4]   2> 	at org.apache.solr.handler.IndexFetcher.fetchLatestIndex(IndexFetcher.java:382)
   [junit4]   2> 	at org.apache.solr.handler.IndexFetcher.fetchLatestIndex(IndexFetcher.java:347)
   [junit4]   2> 	at org.apache.solr.handler.ReplicationHandler.doFetch(ReplicationHandler.java:421)
   [junit4]   2> 	at org.apache.solr.handler.ReplicationHandler.lambda$setupPolling$13(ReplicationHandler.java:1156)
   [junit4]   2> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
   [junit4]   2> 	at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
   [junit4]   2> 	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
   [junit4]   2> 	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
   [junit4]   2> 	at java.lang.Thread.run(Thread.java:748)
   [junit4]   2> 
   [junit4]   2> 80912 WARN  (Thread-1) [    ] o.a.s.c.ZkTestServer Watch limit violations: 
   [junit4]   2> Maximum concurrent create/delete watches above limit:
   [junit4]   2> 
   [junit4]   2> 	22	/solr/aliases.json
   [junit4]   2> 	20	/solr/collections/collection1/terms/shard2
   [junit4]   2> 	14	/solr/collections/collection1/terms/shard1
   [junit4]   2> 	10	/solr/security.json
   [junit4]   2> 	10	/solr/configs/conf1
   [junit4]   2> 	3	/solr/collections/control_collection/terms/shard1
   [junit4]   2> 
   [junit4]   2> Maximum concurrent data watches above limit:
   [junit4]   2> 
   [junit4]   2> 	97	/solr/collections/collection1/state.json
   [junit4]   2> 	22	/solr/clusterprops.json
   [junit4]   2> 	22	/solr/clusterstate.json
   [junit4]   2> 	6	/solr/autoscaling.json
   [junit4]   2> 	4	/solr/overseer_elect/election/72245280566738967-127.0.0.1:45131_y%2Fb-n_0000000004
   [junit4]   2> 	2	/solr/overseer_elect/election/72245280566738959-127.0.0.1:45485_y%2Fb-n_0000000002
   [junit4]   2> 	2	/solr/collections/collection1/leader_elect/shard2/election/72245280566738955-core_node42-n_0000000000
   [junit4]   2> 
   [junit4]   2> Maximum concurrent children watches above limit:
   [junit4]   2> 
   [junit4]   2> 	22	/solr/live_nodes
   [junit4]   2> 	22	/solr/collections
   [junit4]   2> 	6	/solr/autoscaling/events/.scheduled_maintenance
   [junit4]   2> 	6	/solr/autoscaling/events/.auto_add_replicas
   [junit4]   2> 	6	/solr/overseer/collection-queue-work
   [junit4]   2> 	6	/solr/overseer/queue-work
   [junit4]   2> 	5	/solr/overseer/queue
   [junit4]   2> 
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=ChaosMonkeySafeLeaderWithPullReplicasTest -Dtests.method=test -Dtests.seed=809835CCA1EF3CC3 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.badapples=true -Dtests.locale=hu-HU -Dtests.timezone=Africa/Dar_es_Salaam -Dtests.asserts=true -Dtests.file.encoding=UTF-8
   [junit4] FAILURE 77.7s J2 | ChaosMonkeySafeLeaderWithPullReplicasTest.test <<<
   [junit4]    > Throwable #1: java.lang.AssertionError: Timed out waiting for replica core_node52 (1529682826429) to replicate from leader core_node44 (0)
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([809835CCA1EF3CC3:8CC0A160F13513B]:0)
   [junit4]    > 	at org.apache.solr.cloud.AbstractFullDistribZkTestBase.waitForReplicationFromReplicas(AbstractFullDistribZkTestBase.java:2146)
   [junit4]    > 	at org.apache.solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest.test(ChaosMonkeySafeLeaderWithPullReplicasTest.java:211)
   [junit4]    > 	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:993)
   [junit4]    > 	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:968)
   [junit4]    > 	at java.lang.Thread.run(Thread.java:748)
   [junit4]   2> NOTE: leaving temporary files on disk at: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J2/temp/solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest_809835CCA1EF3CC3-001
   [junit4]   2> Jun 22, 2018 3:53:55 PM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
   [junit4]   2> WARNING: Will linger awaiting termination of 1 leaked thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {}, docValues:{}, maxPointsInLeafNode=44, maxMBSortInHeap=7.896967503858964, sim=Asserting(org.apache.lucene.search.similarities.AssertingSimilarity@1c70be), locale=hu-HU, timezone=Africa/Dar_es_Salaam
   [junit4]   2> NOTE: Linux 4.13.0-41-generic i386/Oracle Corporation 1.8.0_172 (32-bit)/cpus=8,threads=1,free=28898776,total=101916672
   [junit4]   2> NOTE: All tests run in this JVM: [ChaosMonkeySafeLeaderWithPullReplicasTest]
   [junit4] Completed [3/5 (1!)] on J2 in 80.27s, 1 test, 1 failure <<< FAILURES!

[...truncated 19 lines...]
BUILD FAILED
/home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/common-build.xml:1568: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/common-build.xml:1092: There were test failures: 5 suites, 5 tests, 1 failure [seed: 809835CCA1EF3CC3]

Total time: 1 minute 53 seconds

[repro] Setting last failure code to 256

[repro] Failures:
[repro]   1/5 failed: org.apache.solr.cloud.ChaosMonkeySafeLeaderWithPullReplicasTest
[repro] Exiting with code 256
+ mv lucene/build lucene/build.repro
+ mv solr/build solr/build.repro
+ mv lucene/build.orig lucene/build
+ mv solr/build.orig solr/build
Archiving artifacts
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
[WARNINGS] Parsing warnings in console log with parser Java Compiler (javac)
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
<Git Blamer> Using GitBlamer to create author and commit information for all warnings.
<Git Blamer> GIT_COMMIT=3b9d3a760a432b97aad2c08b2f778fa2344eb14a, workspace=/var/lib/jenkins/workspace/Lucene-Solr-master-Linux
[WARNINGS] Computing warning deltas based on reference build #22296
Recording test results
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2
Setting ANT_1_8_2_HOME=/var/lib/jenkins/tools/hudson.tasks.Ant_AntInstallation/ANT_1.8.2

[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-10.0.1) - Build # 22301 - Still Unstable!

Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22301/
Java: 64bit/jdk-10.0.1 -XX:-UseCompressedOops -XX:+UseSerialGC

1 tests failed.
FAILED:  org.apache.solr.cloud.cdcr.CdcrBidirectionalTest.testBiDir

Error Message:
Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]

Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
Caused by: java.lang.AssertionError
	at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
	at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
	at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
	at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
	at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 13442 lines...]
   [junit4] Suite: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest
   [junit4]   2> 726810 INFO  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/init-core-data-001
   [junit4]   2> 726813 WARN  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=25 numCloses=25
   [junit4]   2> 726813 INFO  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 726814 INFO  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0)
   [junit4]   2> 726821 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testBiDir
   [junit4]   2> 726821 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 1 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001
   [junit4]   2> 726822 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 726834 INFO  (Thread-2493) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 726834 INFO  (Thread-2493) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 726867 ERROR (Thread-2493) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 726934 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer start zk server on port:38873
   [junit4]   2> 726967 INFO  (zkConnectionManagerCallback-2451-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 726987 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
   [junit4]   2> 727006 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 727006 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 727006 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 727007 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@7c514370{/solr,null,AVAILABLE}
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@55d7ae79{SSL,[ssl, http/1.1]}{127.0.0.1:৩৯৩৪৯}
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.Server Started @৭২৭০৩৩ms
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=39349}
   [junit4]   2> 727008 ERROR (jetty-launcher-2448-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 8.0.0
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T04:08:53.265596Z
   [junit4]   2> 727034 INFO  (zkConnectionManagerCallback-2453-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727035 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 727144 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:38873/solr
   [junit4]   2> 727145 INFO  (zkConnectionManagerCallback-2457-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727146 INFO  (zkConnectionManagerCallback-2459-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727181 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 727182 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:39349_solr
   [junit4]   2> 727182 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.Overseer Overseer (id=72248175661154307-127.0.0.1:39349_solr-n_0000000000) starting
   [junit4]   2> 727186 INFO  (zkConnectionManagerCallback-2466-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727188 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
   [junit4]   2> 727188 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:39349_solr
   [junit4]   2> 727188 INFO  (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727198 INFO  (zkCallback-2465-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727210 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 727225 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727232 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727232 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727233 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/.
   [junit4]   2> 727254 INFO  (zkConnectionManagerCallback-2469-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727258 INFO  (zkConnectionManagerCallback-2472-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727259 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 1 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001
   [junit4]   2> 727260 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 727260 INFO  (Thread-2503) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 727260 INFO  (Thread-2503) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 727267 ERROR (Thread-2503) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 727360 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer start zk server on port:44483
   [junit4]   2> 727361 INFO  (zkConnectionManagerCallback-2476-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727371 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
   [junit4]   2> 727373 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 727373 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 727373 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 727374 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@6ad5a77d{/solr,null,AVAILABLE}
   [junit4]   2> 727374 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@3cfe7871{SSL,[ssl, http/1.1]}{127.0.0.1:৪৩৭৯১}
   [junit4]   2> 727374 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.Server Started @৭২৭৪০০ms
   [junit4]   2> 727374 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=43791}
   [junit4]   2> 727378 ERROR (jetty-launcher-2473-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 8.0.0
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T04:08:53.635561Z
   [junit4]   2> 727380 INFO  (zkConnectionManagerCallback-2478-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727381 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 727532 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44483/solr
   [junit4]   2> 727543 INFO  (zkConnectionManagerCallback-2482-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727546 INFO  (zkConnectionManagerCallback-2484-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727582 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 727582 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:43791_solr
   [junit4]   2> 727583 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.Overseer Overseer (id=72248175688417283-127.0.0.1:43791_solr-n_0000000000) starting
   [junit4]   2> 727590 INFO  (zkConnectionManagerCallback-2491-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727592 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44483/solr ready
   [junit4]   2> 727601 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:43791_solr
   [junit4]   2> 727618 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727623 INFO  (zkCallback-2490-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727632 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 727641 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727646 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727646 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727647 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/.
   [junit4]   2> 727674 INFO  (zkConnectionManagerCallback-2494-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727698 INFO  (zkConnectionManagerCallback-2497-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727703 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.CdcrBidirectionalTest cluster2 zkHost = 127.0.0.1:38873/solr
   [junit4]   2> 727703 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.CdcrBidirectionalTest cluster1 zkHost = 127.0.0.1:44483/solr
   [junit4]   2> 727710 INFO  (zkConnectionManagerCallback-2499-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727722 INFO  (zkConnectionManagerCallback-2503-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727723 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727724 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44483/solr ready
   [junit4]   2> 727738 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=cdcr-cluster1&maxShardsPerNode=2&name=cdcr-cluster1&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 727742 INFO  (OverseerThreadFactory-3404-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection cdcr-cluster1
   [junit4]   2> 727847 INFO  (OverseerStateUpdate-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster1",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"cdcr-cluster1_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:43791/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 727850 INFO  (OverseerStateUpdate-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster1",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"cdcr-cluster1_shard2_replica_n3",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:43791/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 728103 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr    x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster1_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 728104 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr    x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 728103 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr    x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster1_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 729144 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 729149 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.IndexSchema [cdcr-cluster1_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 729150 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 729150 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster1_shard1_replica_n1' using configuration from collection cdcr-cluster1, trusted=true
   [junit4]   2> 729151 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 729152 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.core.cdcr-cluster1.shard1.replica_n1' (registry 'solr.core.cdcr-cluster1.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 729154 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 729156 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [[cdcr-cluster1_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/cdcr-cluster1_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/./cdcr-cluster1_shard1_replica_n1/data/]
   [junit4]   2> 729156 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.IndexSchema [cdcr-cluster1_shard2_replica_n3] Schema name=minimal
   [junit4]   2> 729157 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 729157 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster1_shard2_replica_n3' using configuration from collection cdcr-cluster1, trusted=true
   [junit4]   2> 729158 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.core.cdcr-cluster1.shard2.replica_n3' (registry 'solr.core.cdcr-cluster1.shard2.replica_n3') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 729158 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 729158 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [[cdcr-cluster1_shard2_replica_n3] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/cdcr-cluster1_shard2_replica_n3], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/./cdcr-cluster1_shard2_replica_n3/data/]
   [junit4]   2> 729331 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 729331 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 729331 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 729331 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 729332 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 729332 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 729332 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 729332 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 729333 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@4cade66a[cdcr-cluster1_shard1_replica_n1] main]
   [junit4]   2> 729333 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@534a2f15[cdcr-cluster1_shard2_replica_n3] main]
   [junit4]   2> 729334 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster1
   [junit4]   2> 729334 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster1
   [junit4]   2> 729334 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster1
   [junit4]   2> 729334 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster1
   [junit4]   2> 729334 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster1/managed-schema
   [junit4]   2> 729334 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster1/managed-schema
   [junit4]   2> 729334 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 729334 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 729335 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 729335 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 729338 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.CdcrBufferStateManager Created znode /collections/cdcr-cluster1/cdcr/state/buffer
   [junit4]   2> 729340 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrProcessStateManager Created znode /collections/cdcr-cluster1/cdcr/state/process
   [junit4]   2> 729353 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034951234191360
   [junit4]   2> 729353 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034951234191360
   [junit4]   2> 729354 INFO  (searcherExecutor-3410-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard2_replica_n3 c:cdcr-cluster1 s:shard2 r:core_node4) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster1_shard2_replica_n3] Registered new searcher Searcher@534a2f15[cdcr-cluster1_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 729354 INFO  (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@4cade66a[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 729357 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard1 to Terms{values={core_node2=0}, version=0}
   [junit4]   2> 729358 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard2 to Terms{values={core_node4=0}, version=0}
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/ has no replicas
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/ has no replicas
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 729361 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster1:shard1
   [junit4]   2> 729364 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster1:shard2
   [junit4]   2> 729364 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/ shard1
   [junit4]   2> 729366 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/ shard2
   [junit4]   2> 729476 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 729477 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster1_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1374
   [junit4]   2> 729516 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 729517 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster1_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1414
   [junit4]   2> 729519 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 729618 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/state.json] for collection [cdcr-cluster1] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 729744 INFO  (OverseerCollectionConfigSetProcessor-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 730519 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=cdcr-cluster1&maxShardsPerNode=2&name=cdcr-cluster1&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2} status=0 QTime=2781
   [junit4]   2> 730538 INFO  (zkConnectionManagerCallback-2507-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 730545 INFO  (zkConnectionManagerCallback-2511-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 730546 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 730560 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
   [junit4]   2> 730623 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=cdcr-cluster2&maxShardsPerNode=2&name=cdcr-cluster2&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 730639 INFO  (OverseerThreadFactory-3391-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection cdcr-cluster2
   [junit4]   2> 730749 INFO  (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster2",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"cdcr-cluster2_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:39349/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 730751 INFO  (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster2",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"cdcr-cluster2_shard2_replica_n3",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:39349/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 730989 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr    x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster2_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 730990 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr    x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 731031 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr    x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster2_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 732008 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 732013 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.IndexSchema [cdcr-cluster2_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 732043 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 732044 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster2_shard1_replica_n1' using configuration from collection cdcr-cluster2, trusted=true
   [junit4]   2> 732044 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.core.cdcr-cluster2.shard1.replica_n1' (registry 'solr.core.cdcr-cluster2.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 732044 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 732044 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [[cdcr-cluster2_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/cdcr-cluster2_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/./cdcr-cluster2_shard1_replica_n1/data/]
   [junit4]   2> 732055 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 732063 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.IndexSchema [cdcr-cluster2_shard2_replica_n3] Schema name=minimal
   [junit4]   2> 732064 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 732064 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster2_shard2_replica_n3' using configuration from collection cdcr-cluster2, trusted=true
   [junit4]   2> 732065 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.core.cdcr-cluster2.shard2.replica_n3' (registry 'solr.core.cdcr-cluster2.shard2.replica_n3') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 732065 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 732065 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [[cdcr-cluster2_shard2_replica_n3] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/cdcr-cluster2_shard2_replica_n3], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/./cdcr-cluster2_shard2_replica_n3/data/]
   [junit4]   2> 732150 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 732151 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 732152 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 732152 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 732155 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@116811f2[cdcr-cluster2_shard1_replica_n1] main]
   [junit4]   2> 732156 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster2
   [junit4]   2> 732156 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster2
   [junit4]   2> 732156 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster2/managed-schema
   [junit4]   2> 732156 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 732157 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 732159 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrBufferStateManager Created znode /collections/cdcr-cluster2/cdcr/state/buffer
   [junit4]   2> 732165 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrProcessStateManager Created znode /collections/cdcr-cluster2/cdcr/state/process
   [junit4]   2> 732206 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034954225778688
   [junit4]   2> 732208 INFO  (searcherExecutor-3421-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster2_shard1_replica_n1] Registered new searcher Searcher@116811f2[cdcr-cluster2_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 732208 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 732208 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 732209 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 732209 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 732211 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster2/terms/shard1 to Terms{values={core_node2=0}, version=0}
   [junit4]   2> 732212 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@7910e119[cdcr-cluster2_shard2_replica_n3] main]
   [junit4]   2> 732212 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster2
   [junit4]   2> 732213 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster2
   [junit4]   2> 732213 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster2/managed-schema
   [junit4]   2> 732213 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 732214 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/ has no replicas
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 732216 INFO  (zkCallback-2458-thread-2) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster2:shard1
   [junit4]   2> 732217 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/ shard1
   [junit4]   2> 732224 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034954244653056
   [junit4]   2> 732252 INFO  (searcherExecutor-3422-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster2_shard2_replica_n3] Registered new searcher Searcher@7910e119[cdcr-cluster2_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 732253 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster2/terms/shard2 to Terms{values={core_node4=0}, version=0}
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/ has no replicas
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 732256 INFO  (zkCallback-2458-thread-2) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster2:shard2
   [junit4]   2> 732258 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/ shard2
   [junit4]   2> 732360 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 732361 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster2_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1330
   [junit4]   2> 732368 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 732369 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster2_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1380
   [junit4]   2> 732373 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 732470 INFO  (zkCallback-2458-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster2/state.json] for collection [cdcr-cluster2] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 732662 INFO  (OverseerCollectionConfigSetProcessor-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 733373 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=cdcr-cluster2&maxShardsPerNode=2&name=cdcr-cluster2&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2} status=0 QTime=2750
   [junit4]   2> 733390 INFO  (zkConnectionManagerCallback-2518-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 733391 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 733391 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
   [junit4]   2> 733421 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 733421 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=3
   [junit4]   2> 733443 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 733443 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733444 INFO  (qtp185556678-10496) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=45
   [junit4]   2> 733445 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard1
   [junit4]   2> 733445 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2, shard: shard1
   [junit4]   2> 733447 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Submitting bootstrap task to executor
   [junit4]   2> 733451 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2 shard: shard1 leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/
   [junit4]   2> 733461 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrProcessStateManager The CDCR process state has changed: WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/cdcr/state/process @ cdcr-cluster1:shard2
   [junit4]   2> 733461 INFO  (zkCallback-2483-thread-2) [    ] o.a.s.h.CdcrProcessStateManager The CDCR process state has changed: WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/cdcr/state/process @ cdcr-cluster1:shard1
   [junit4]   2> 733462 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrProcessStateManager Received new CDCR process state from watcher: STARTED @ cdcr-cluster1:shard2
   [junit4]   2> 733462 INFO  (zkCallback-2483-thread-2) [    ] o.a.s.h.CdcrProcessStateManager Received new CDCR process state from watcher: STARTED @ cdcr-cluster1:shard1
   [junit4]   2> 733462 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/cdcr params={qt=/cdcr&_stateVer_=cdcr-cluster1:4&action=start&wt=javabin&version=2} status=0 QTime=85
   [junit4]   2> 733514 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={qt=/cdcr&masterUrl=https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/&action=BOOTSTRAP&wt=javabin&version=2} status=0 QTime=5
   [junit4]   2> 733516 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733518 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager CDCR bootstrap running for 1 seconds, sleeping for 2000 ms
   [junit4]   2> 733526 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 733534 INFO  (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034955618287616,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 733542 INFO  (zkConnectionManagerCallback-2523-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 733542 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 733543 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
   [junit4]   2> 733554 INFO  (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 733555 INFO  (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 733555 INFO  (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={waitSearcher=true&openSearcher=false&commit=true&softCommit=false&commit_end_point=true&wt=javabin&version=2} status=0 QTime=22
   [junit4]   2> 733585 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 733585 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733609 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 733609 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733610 INFO  (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=64
   [junit4]   2> 733610 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard2
   [junit4]   2> 733610 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2, shard: shard2
   [junit4]   2> 733610 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Submitting bootstrap task to executor
   [junit4]   2> 733615 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2 shard: shard2 leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/
   [junit4]   2> 733622 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/replication params={qt=/replication&wt=javabin&version=2&command=indexversion} status=0 QTime=0
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Master's generation: 1
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Master's version: 0
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Slave's generation: 1
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Slave's version: 0
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher New index in Master. Deleting mine...
   [junit4]   2> 733627 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@860e11[cdcr-cluster2_shard1_replica_n1] main]
   [junit4]   2> 733628 INFO  (searcherExecutor-3421-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster2_shard1_replica_n1] Registered new searcher Searcher@860e11[cdcr-cluster2_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 733628 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler No replay needed.
   [junit4]   2> 733628 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={qt=/cdcr&masterUrl=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&action=BOOTSTRAP&wt=javabin&version=2} status=0 QTime=9
   [junit4]   2> 733629 INFO  (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733630 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager CDCR bootstrap running for 1 seconds, sleeping for 2000 ms
   [junit4]   2> 733630 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 733641 INFO  (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034955730485248,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 733646 INFO  (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 733646 INFO  (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 733646 INFO  (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={waitSearcher=true&openSearcher=false&commit=true&softCommit=false&commit_end_point=true&wt=javabin&version=2} status=0 QTime=6
   [junit4]   2> 733650 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/replication params={qt=/replication&wt=javabin&version=2&command=indexversion} status=0 QTime=0
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Master's generation: 1
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Master's version: 0
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Slave's generation: 1
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Slave's version: 0
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher New index in Master. Deleting mine...
   [junit4]   2> 733651 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@3853640e[cdcr-cluster2_shard2_replica_n3] main]
   [junit4]   2> 733652 INFO  (searcherExecutor-3422-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster2_shard2_replica_n3] Registered new searcher Searcher@3853640e[cdcr-cluster2_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 733652 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler No replay needed.
   [junit4]   2> 735463 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.CdcrBidirectionalTest Adding 10 docs with commit=true, numDocs=100
   [junit4]   2> 735469 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard1 to Terms{values={core_node2=1}, version=1}
   [junit4]   2> 735469 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=3
   [junit4]   2> 735469 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard2 to Terms{values={core_node4=1}, version=1}
   [junit4]   2> 735469 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=5
   [junit4]   2> 735474 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957652525056,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 735474 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@62bc2cb1 commitCommandVersion:1604034957652525056
   [junit4]   2> 735478 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957656719360,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 735478 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@6850d084 commitCommandVersion:1604034957656719360
   [junit4]   2> 735518 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735519 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager CDCR bootstrap successful in 3 seconds
   [junit4]   2> 735523 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 735523 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735525 INFO  (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 735525 INFO  (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735528 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=9
   [junit4]   2> 735529 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard1
   [junit4]   2> 735529 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Bootstrap successful, giving the go-ahead to replicator
   [junit4]   2> 735630 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735631 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager CDCR bootstrap successful in 3 seconds
   [junit4]   2> 735637 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 735637 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 735637 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735637 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735638 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=6
   [junit4]   2> 735638 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard2
   [junit4]   2> 735638 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Bootstrap successful, giving the go-ahead to replicator
   [junit4]   2> 735653 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@601f19c[cdcr-cluster1_shard1_replica_n1] main]
   [junit4]   2> 735654 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 735656 INFO  (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@601f19c[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C43)))}
   [junit4]   2> 735656 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false} status=0 QTime=182
   [junit4]   2> 735667 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@6fb4ca9f[cdcr-cluster1_shard2_replica_n3] main]
   [junit4]   2> 735668 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 735669 INFO  (searcherExecutor-3410-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard2_replica_n3 c:cdcr-cluster1 s:shard2 r:core_node4) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster1_shard2_replica_n3] Registered new searcher Searcher@6fb4ca9f[cdcr-cluster1_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C57)))}
   [junit4]   2> 735669 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false} status=0 QTime=191
   [junit4]   2> 735669 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&waitSearcher=true&commit=true&softCommit=false&wt=javabin&version=2} status=0 QTime=199
   [junit4]   2> 735670 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.CdcrBidirectionalTest Adding 10 docs with commit=true, numDocs=200
   [junit4]   2> 735680 INFO  (qtp804107552-10610) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=9
   [junit4]   2> 735683 INFO  (qtp804107552-10609) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=6
   [junit4]   2> 735684 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957872726016,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 735684 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@6850d084 commitCommandVersion:1604034957872726016
   [junit4]   2> 735686 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957874823168,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 735686 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@62bc2cb1 commitCommandVersion:1604034957874823168
   [junit4]   2> 735972 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@3f3de49e[cdcr-cluster1_shard1_replica_n1] main]
   [junit4]   2> 735973 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 735974 INFO  (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@3f3de49e[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C43) Uninverting(_1(8.0.0):C54)))}
   [junit4]   2> 735974 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}

[...truncated too long message...]

try=solr.collection.cdcr-cluster2.shard2.leader, tag=18455e69
   [junit4]   2> 760651 INFO  (coreCloseExecutor-3451-thread-2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Solr core is being closed - shutting down CDCR handler @ cdcr-cluster2:shard2
   [junit4]   2> 760731 INFO  (jetty-closer-2449-thread-1) [    ] o.a.s.c.Overseer Overseer (id=72248175661154307-127.0.0.1:39349_solr-n_0000000000) closing
   [junit4]   2> 760731 INFO  (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.Overseer Overseer Loop exiting : 127.0.0.1:39349_solr
   [junit4]   2> 760731 WARN  (OverseerAutoScalingTriggerThread-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread woken up but we are closed, exiting.
   [junit4]   2> 760732 INFO  (jetty-closer-2449-thread-1) [    ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@7c514370{/solr,null,UNAVAILABLE}
   [junit4]   2> 760732 INFO  (jetty-closer-2449-thread-1) [    ] o.e.j.s.session node0 Stopped scavenging
   [junit4]   2> 760733 ERROR (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 760733 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:38873 38873
   [junit4]   2> 760778 INFO  (Thread-2493) [    ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:38873 38873
   [junit4]   2> 760778 WARN  (Thread-2493) [    ] o.a.s.c.ZkTestServer Watch limit violations: 
   [junit4]   2> Maximum concurrent create/delete watches above limit:
   [junit4]   2> 
   [junit4]   2> 	5	/solr/aliases.json
   [junit4]   2> 	3	/solr/collections/cdcr-cluster2/terms/shard2
   [junit4]   2> 	3	/solr/collections/cdcr-cluster2/terms/shard1
   [junit4]   2> 	2	/solr/configs/cdcr-cluster2/managed-schema
   [junit4]   2> 
   [junit4]   2> Maximum concurrent data watches above limit:
   [junit4]   2> 
   [junit4]   2> 	6	/solr/collections/cdcr-cluster2/cdcr/state/process
   [junit4]   2> 	5	/solr/clusterstate.json
   [junit4]   2> 	5	/solr/clusterprops.json
   [junit4]   2> 	3	/solr/collections/cdcr-cluster2/state.json
   [junit4]   2> 	2	/solr/collections/cdcr-cluster2/cdcr/state/buffer
   [junit4]   2> 
   [junit4]   2> Maximum concurrent children watches above limit:
   [junit4]   2> 
   [junit4]   2> 	5	/solr/live_nodes
   [junit4]   2> 	5	/solr/collections
   [junit4]   2> 
   [junit4]   2> 760779 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.SolrTestCaseJ4 ###Ending testBiDir
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=CdcrBidirectionalTest -Dtests.method=testBiDir -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8
   [junit4] ERROR   34.0s J0 | CdcrBidirectionalTest.testBiDir <<<
   [junit4]    > Throwable #1: com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
   [junit4]    > Caused by: java.lang.AssertionError
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
   [junit4]    > 	at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
   [junit4]    > 	at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
   [junit4]    > 	at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
   [junit4]    > 	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
   [junit4]    > 	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
   [junit4]    > 	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
   [junit4]    > 	at java.base/java.lang.Thread.run(Thread.java:844)Throwable #2: com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10729, name=cdcr-replicator-3446-thread-2, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
   [junit4]    > Caused by: java.lang.AssertionError
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
   [junit4]    > 	at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
   [junit4]    > 	at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
   [junit4]    > 	at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
   [junit4]    > 	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
   [junit4]    > 	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
   [junit4]    > 	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
   [junit4]    > 	at java.base/java.lang.Thread.run(Thread.java:844)
   [junit4]   2> 760780 INFO  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 ------------------------------------------------------- Done waiting for tracked resources to be released
   [junit4]   2> Jun 23, 2018 4:09:27 AM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
   [junit4]   2> WARNING: Will linger awaiting termination of 5 leaked thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene70), sim=Asserting(org.apache.lucene.search.similarities.AssertingSimilarity@8d81519), locale=bn-IN, timezone=Asia/Tomsk
   [junit4]   2> NOTE: Linux 4.13.0-41-generic amd64/Oracle Corporation 10.0.1 (64-bit)/cpus=8,threads=1,free=149515200,total=518979584
   [junit4]   2> NOTE: All tests run in this JVM: [TestNestedDocsSort, TestMaxScoreQueryParser, URLClassifyProcessorTest, ZkCLITest, TestHttpShardHandlerFactory, SimpleCollectionCreateDeleteTest, TestClusterStateMutator, TestMinMaxOnMultiValuedField, ZkNodePropsTest, TestSort, TestComplexPhraseLeadingWildcard, TestHdfsBackupRestoreCore, TestSystemIdResolver, CursorPagingTest, TestSolrCloudWithDelegationTokens, ChaosMonkeyNothingIsSafeTest, TestBulkSchemaConcurrent, TestCollationFieldDocValues, DeleteLastCustomShardedReplicaTest, TestNumericRangeQuery32, TestRTGBase, TestManagedSchemaAPI, FileUtilsTest, BlockJoinFacetSimpleTest, TestShortCircuitedRequests, TestJettySolrRunner, BigEndianAscendingWordSerializerTest, TestCSVLoader, TestExecutePlanAction, StatsComponentTest, AtomicUpdateProcessorFactoryTest, RegexBytesRefFilterTest, HdfsUnloadDistributedZkTest, HealthCheckHandlerTest, SolrCoreMetricManagerTest, TestRequestStatusCollectionAPI, BadCopyFieldTest, SolrXmlInZkTest, TestRemoteStreaming, IndexBasedSpellCheckerTest, TestStressCloudBlindAtomicUpdates, TestPrepRecovery, TestUtilizeNode, RequestHandlersTest, TestObjectReleaseTracker, TestDistribIDF, HdfsAutoAddReplicasIntegrationTest, SchemaApiFailureTest, FieldAnalysisRequestHandlerTest, SystemInfoHandlerTest, DistributedQueryElevationComponentTest, BJQParserTest, MBeansHandlerTest, CollectionsAPIAsyncDistributedZkTest, DocumentAnalysisRequestHandlerTest, AutoScalingHandlerTest, PKIAuthenticationIntegrationTest, TestRawTransformer, XmlInterpolationTest, TestSolrDeletionPolicy1, TestExportWriter, TestManagedStopFilterFactory, TestQuerySenderNoQuery, MetricsHistoryHandlerTest, TestDocBasedVersionConstraints, NodeMarkersRegistrationTest, FullSolrCloudDistribCmdsTest, TestCloudDeleteByQuery, TestSolrCLIRunExample, ShufflingReplicaListTransformerTest, AssignTest, BlobRepositoryCloudTest, TestCrossCoreJoin, HDFSCollectionsAPITest, HttpSolrCallGetCoreTest, TestLeaderElectionWithEmptyReplica, TestLuceneMatchVersion, TestLegacyNumericUtils, HdfsChaosMonkeyNothingIsSafeTest, ConfigureRecoveryStrategyTest, TestIBSimilarityFactory, TestEmbeddedSolrServerAdminHandler, HdfsSyncSliceTest, CoreMergeIndexesAdminHandlerTest, CustomCollectionTest, TestConfig, SimpleFacetsTest, OverseerStatusTest, HighlighterConfigTest, RollingRestartTest, TestSolrQueryParser, TestDocTermOrdsUninvertLimit, TestDistributedMap, SolrCloudReportersTest, TestXmlQParser, TestIndexSearcher, TestGroupingSearch, TestCustomStream, SharedFSAutoReplicaFailoverTest, RAMDirectoryFactoryTest, TestJsonRequest, ProtectedTermFilterFactoryTest, TestJsonFacetRefinement, CdcrBidirectionalTest]
   [junit4] Completed [310/821 (1!)] on J0 in 34.51s, 1 test, 1 error <<< FAILURES!

[...truncated 45707 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22301/consoleText

[repro] Revision: 3b9d3a760a432b97aad2c08b2f778fa2344eb14a

[repro] Ant options: "-Dargs=-XX:-UseCompressedOops -XX:+UseSerialGC"
[repro] Repro line:  ant test  -Dtestcase=CdcrBidirectionalTest -Dtests.method=testBiDir -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8

[repro] ant clean

[...truncated 6 lines...]
[repro] Test suites by module:
[repro]    solr/core
[repro]       CdcrBidirectionalTest
[repro] ant compile-test

[...truncated 3305 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.CdcrBidirectionalTest" -Dtests.showOutput=onerror "-Dargs=-XX:-UseCompressedOops -XX:+UseSerialGC" -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8

[...truncated 76 lines...]
[repro] Failures:
[repro]   0/5 failed: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest
[repro] Exiting with code 0

[...truncated 40 lines...]

[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-10) - Build # 22300 - Still Unstable!

Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22300/
Java: 64bit/jdk-10 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest.testSimple

Error Message:
Waiting for collection testSimple2 null Live Nodes: [127.0.0.1:34223_solr, 127.0.0.1:45039_solr] Last available state: DocCollection(testSimple2//collections/testSimple2/state.json/21)={   "pullReplicas":"0",   "replicationFactor":"2",   "shards":{     "shard1":{       "range":"80000000-ffffffff",       "state":"active",       "replicas":{         "core_node3":{           "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/",           "base_url":"http://127.0.0.1:34223/solr",           "node_name":"127.0.0.1:34223_solr",           "type":"NRT",           "force_set_state":"false",           "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/tlog",           "core":"testSimple2_shard1_replica_n1",           "shared_storage":"true",           "state":"active",           "leader":"true"},         "core_node5":{           "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/",           "base_url":"http://127.0.0.1:39879/solr",           "node_name":"127.0.0.1:39879_solr",           "type":"NRT",           "force_set_state":"false",           "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/tlog",           "core":"testSimple2_shard1_replica_n2",           "shared_storage":"true",           "state":"down"}}},     "shard2":{       "range":"0-7fffffff",       "state":"active",       "replicas":{         "core_node6":{           "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/",           "base_url":"http://127.0.0.1:34223/solr",           "node_name":"127.0.0.1:34223_solr",           "type":"NRT",           "force_set_state":"false",           "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/tlog",           "core":"testSimple2_shard2_replica_n4",           "shared_storage":"true",           "state":"active",           "leader":"true"},         "core_node8":{           "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/",           "base_url":"http://127.0.0.1:39879/solr",           "node_name":"127.0.0.1:39879_solr",           "type":"NRT",           "force_set_state":"false",           "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/tlog",           "core":"testSimple2_shard2_replica_n7",           "shared_storage":"true",           "state":"down"}}}},   "router":{"name":"compositeId"},   "maxShardsPerNode":"2",   "autoAddReplicas":"true",   "nrtReplicas":"2",   "tlogReplicas":"0"}

Stack Trace:
java.lang.AssertionError: Waiting for collection testSimple2
null
Live Nodes: [127.0.0.1:34223_solr, 127.0.0.1:45039_solr]
Last available state: DocCollection(testSimple2//collections/testSimple2/state.json/21)={
  "pullReplicas":"0",
  "replicationFactor":"2",
  "shards":{
    "shard1":{
      "range":"80000000-ffffffff",
      "state":"active",
      "replicas":{
        "core_node3":{
          "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/",
          "base_url":"http://127.0.0.1:34223/solr",
          "node_name":"127.0.0.1:34223_solr",
          "type":"NRT",
          "force_set_state":"false",
          "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/tlog",
          "core":"testSimple2_shard1_replica_n1",
          "shared_storage":"true",
          "state":"active",
          "leader":"true"},
        "core_node5":{
          "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/",
          "base_url":"http://127.0.0.1:39879/solr",
          "node_name":"127.0.0.1:39879_solr",
          "type":"NRT",
          "force_set_state":"false",
          "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/tlog",
          "core":"testSimple2_shard1_replica_n2",
          "shared_storage":"true",
          "state":"down"}}},
    "shard2":{
      "range":"0-7fffffff",
      "state":"active",
      "replicas":{
        "core_node6":{
          "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/",
          "base_url":"http://127.0.0.1:34223/solr",
          "node_name":"127.0.0.1:34223_solr",
          "type":"NRT",
          "force_set_state":"false",
          "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/tlog",
          "core":"testSimple2_shard2_replica_n4",
          "shared_storage":"true",
          "state":"active",
          "leader":"true"},
        "core_node8":{
          "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/",
          "base_url":"http://127.0.0.1:39879/solr",
          "node_name":"127.0.0.1:39879_solr",
          "type":"NRT",
          "force_set_state":"false",
          "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/tlog",
          "core":"testSimple2_shard2_replica_n7",
          "shared_storage":"true",
          "state":"down"}}}},
  "router":{"name":"compositeId"},
  "maxShardsPerNode":"2",
  "autoAddReplicas":"true",
  "nrtReplicas":"2",
  "tlogReplicas":"0"}
	at __randomizedtesting.SeedInfo.seed([A483062C73879C9B:9C3022D25474484A]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.apache.solr.cloud.SolrCloudTestCase.waitForState(SolrCloudTestCase.java:278)
	at org.apache.solr.cloud.autoscaling.AutoAddReplicasIntegrationTest.testSimple(AutoAddReplicasIntegrationTest.java:129)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.base/java.lang.reflect.Method.invoke(Method.java:564)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 1871 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J0-20180622_233120_8438624310112548928347.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J2-20180622_233120_843109218256741183964.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J1-20180622_233120_84215878213975654336568.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 298 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J1-20180622_234003_82713161516520257880618.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J2-20180622_234003_82211985366498351429167.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J0-20180622_234003_8187629202302504117049.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1080 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J0-20180622_234142_9048161528471907446887.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J2-20180622_234142_90416258816079129641780.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J1-20180622_234142_90412005419182290644372.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 258 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J0-20180622_234332_4914664188095482509082.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J2-20180622_234332_4922443722992351897655.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J1-20180622_234332_4914364283643442390734.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 250 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J0-20180622_234345_5569241898986794068605.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J2-20180622_234345_55612839679531665619473.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J1-20180622_234345_55614100303459436628584.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 162 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J1-20180622_234408_1182183404860930597018.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J0-20180622_234408_11817221358498753266703.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J2-20180622_234408_1191547639190121044076.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 210 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J0-20180622_234412_40714371405619675624954.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J2-20180622_234412_4078848850315630926441.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J1-20180622_234412_4079975389867110683768.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 174 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J1-20180622_234421_51010637845793649127342.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J0-20180622_234421_5103569472874704098846.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J2-20180622_234421_51016436102187679443010.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 176 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J2-20180622_234424_31610752690720732553073.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J0-20180622_234424_3163920468419256368342.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J1-20180622_234424_31613507803586776163227.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 161 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/smartcn/test/temp/junit4-J0-20180622_234434_0177751930883938230811.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/smartcn/test/temp/junit4-J1-20180622_234434_01715792206888438466742.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 162 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J1-20180622_234443_3098926868024307710344.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J2-20180622_234443_30910390262105719088600.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J0-20180622_234443_3078828821911177442895.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 177 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/uima/test/temp/junit4-J0-20180622_234446_5954096128335448572284.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 153 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J1-20180622_234455_65114409003380652460182.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J2-20180622_234455_6514623326656775997862.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 19 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J0-20180622_234455_6517020648835771442894.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1391 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J2-20180622_234519_1932894334289275466381.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J1-20180622_234519_1932449911797630396956.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J0-20180622_234519_1936352665634460400351.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 253 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J2-20180622_234527_4918016004016563548263.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J0-20180622_234527_49118385300419579868577.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J1-20180622_234527_4915089559484994378250.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 262 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J1-20180622_234539_34511673298149119415382.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J0-20180622_234539_3442056690456979073956.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J2-20180622_234539_3456681254644361673058.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 231 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J1-20180622_234754_3966129908289707100345.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J2-20180622_234754_39610786966458196672386.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J0-20180622_234754_39611330166380200542034.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 176 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J1-20180622_234757_3282744005959643641695.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J2-20180622_234757_328184188593405154770.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J0-20180622_234757_32813663791879804413349.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 235 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J1-20180622_234803_6451920093429510786067.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J2-20180622_234803_6456139679212212410020.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J0-20180622_234803_64513176520531889858516.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 185 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J2-20180622_234842_66210931058644992560173.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J1-20180622_234842_66216254532774656442410.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J0-20180622_234842_6625491280472342315484.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 255 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J1-20180622_234900_48815603692318856114647.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J2-20180622_234900_48816107964350347376247.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J0-20180622_234900_48816853337107767940056.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 163 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J0-20180622_234916_9509003729064799529499.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J2-20180622_234916_9509833749328796722833.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J1-20180622_234916_9501938938754275132751.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 155 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/memory/test/temp/junit4-J0-20180622_234945_85111956510346757286560.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/memory/test/temp/junit4-J1-20180622_234945_8526925569142241457753.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 189 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J2-20180622_234957_89716020023611267993152.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J1-20180622_234957_89713374075461271625175.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J0-20180622_234957_897995069218042617454.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 304 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J1-20180622_235015_07816445064399041937754.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J2-20180622_235015_07817300623641250331463.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J0-20180622_235015_0777051079486333859241.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 229 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J2-20180622_235023_9761355228138200064238.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J0-20180622_235023_9765821768678419074163.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J1-20180622_235023_9764016929074376670969.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 212 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J1-20180622_235031_70213674486619839082375.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J0-20180622_235031_70211591443367436951168.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J2-20180622_235031_70217646439452202035793.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 227 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J0-20180622_235044_36518075386306318393681.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J2-20180622_235044_3655371863721534862338.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 7 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J1-20180622_235044_3653195148306886251101.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 298 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J0-20180622_235133_51311242690447766539259.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J2-20180622_235133_5131023607900679846491.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J1-20180622_235133_5137580074882706317880.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 153 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial/test/temp/junit4-J0-20180622_235158_4753421684692232253190.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 181 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J0-20180622_235200_23017643266495511437138.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J1-20180622_235200_2306779390704051548979.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J2-20180622_235200_23015255487659489383656.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 264 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J2-20180622_235320_44513781554556465979598.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J0-20180622_235320_4444365517420477492284.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J1-20180622_235320_4449138489014132900971.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3929 lines...]
   [junit4] Suite: org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest
   [junit4]   2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/init-core-data-001
   [junit4]   2> 1628046 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=1 numCloses=1
   [junit4]   2> 1628046 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.SolrTestCaseJ4 Using TrieFields (NUMERIC_POINTS_SYSPROP=false) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 1628047 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0)
   [junit4]   2> 1628047 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 1628047 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 3 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001
   [junit4]   2> 1628047 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 1628048 INFO  (Thread-4546) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 1628048 INFO  (Thread-4546) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 1628049 ERROR (Thread-4546) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 1628148 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.ZkTestServer start zk server on port:36131
   [junit4]   2> 1628149 INFO  (zkConnectionManagerCallback-10100-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@4ebf9fd8{/solr,null,AVAILABLE}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@636e248d{/solr,null,AVAILABLE}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@3a13410e{/solr,null,AVAILABLE}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.AbstractConnector Started ServerConnector@295b2198{HTTP/1.1,[http/1.1]}{127.0.0.1:45039}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@dd199bf{HTTP/1.1,[http/1.1]}{127.0.0.1:39879}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.Server Started @1628177ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.AbstractConnector Started ServerConnector@7ef44222{HTTP/1.1,[http/1.1]}{127.0.0.1:34223}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.Server Started @1628177ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.Server Started @1628177ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=45039}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=39879}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=34223}
   [junit4]   2> 1628152 ERROR (jetty-launcher-10097-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1628152 ERROR (jetty-launcher-10097-thread-3) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1628152 ERROR (jetty-launcher-10097-thread-2) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 8.0.0
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 8.0.0
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 8.0.0
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T00:21:31.684042Z
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T00:21:31.684074Z
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T00:21:31.684084Z
   [junit4]   2> 1628153 INFO  (zkConnectionManagerCallback-10104-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628153 INFO  (zkConnectionManagerCallback-10106-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628153 INFO  (zkConnectionManagerCallback-10103-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628154 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 1628154 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 1628154 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1628156 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1628156 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1628157 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1628243 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
   [junit4]   2> 1628244 INFO  (zkConnectionManagerCallback-10110-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628245 INFO  (zkConnectionManagerCallback-10112-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628251 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
   [junit4]   2> 1628260 INFO  (zkConnectionManagerCallback-10118-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628265 INFO  (zkConnectionManagerCallback-10120-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628269 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1628269 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:45039_solr
   [junit4]   2> 1628270 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.Overseer Overseer (id=72247281650171911-127.0.0.1:45039_solr-n_0000000000) starting
   [junit4]   2> 1628281 INFO  (zkConnectionManagerCallback-10127-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628282 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
   [junit4]   2> 1628284 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:45039_solr
   [junit4]   2> 1628285 INFO  (zkCallback-10126-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1628285 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Adding .auto_add_replicas and .scheduled_maintenance triggers
   [junit4]   2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 1
   [junit4]   2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 1, lastZnodeVersion -1
   [junit4]   2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 1
   [junit4]   2> 1628289 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:45039_solr]
   [junit4]   2> 1628289 DEBUG (ScheduledTrigger-8462-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 1628292 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1628296 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 1628296 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 1, lastZnodeVersion 1
   [junit4]   2> 1628297 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1628297 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1628298 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1628298 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:39879_solr
   [junit4]   2> 1628299 INFO  (zkCallback-10119-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1628299 INFO  (zkCallback-10126-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1628305 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1628313 INFO  (zkCallback-10111-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1628317 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628317 INFO  (zkConnectionManagerCallback-10133-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628318 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 1628318 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
   [junit4]   2> 1628318 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1628323 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628323 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628324 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node3/.
   [junit4]   2> 1628330 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628336 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628336 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628337 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node1/.
   [junit4]   2> 1628408 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
   [junit4]   2> 1628409 INFO  (zkConnectionManagerCallback-10138-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628410 INFO  (zkConnectionManagerCallback-10140-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628413 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 1628414 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1628415 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1628415 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34223_solr
   [junit4]   2> 1628415 INFO  (zkCallback-10119-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628415 INFO  (zkCallback-10126-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628415 INFO  (zkCallback-10111-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628415 INFO  (zkCallback-10139-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628415 INFO  (zkCallback-10132-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628424 INFO  (zkConnectionManagerCallback-10147-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628425 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 1628425 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
   [junit4]   2> 1628426 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1628435 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628440 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628440 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628441 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node2/.
   [junit4]   2> 1628457 INFO  (zkConnectionManagerCallback-10150-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628459 INFO  (zkConnectionManagerCallback-10155-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628459 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 1628460 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 1628518 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 1628524 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 1628525 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 1628531 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_localdomain_33425_hdfs____.4bsgsv/webapp
   [junit4]   2> 1628601 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:33425
   [junit4]   2> 1628678 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 1628679 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 1628683 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_34005_datanode____3r5l3v/webapp
   [junit4]   2> 1628761 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:34005
   [junit4]   2> 1628785 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 1628786 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 1628789 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_32985_datanode____.yy7xwi/webapp
   [junit4]   2> 1628864 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data2/]]  heartbeating to localhost.localdomain/127.0.0.1:46029) [    ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
   [junit4]   2> 1628869 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0xa4871f7444f09: from storage DS-53fc62e1-2ec1-4aa4-a0a1-da327bd2c668 node DatanodeRegistration(127.0.0.1:40969, datanodeUuid=6fa86369-873f-4f5b-b41f-1e017be7aac9, infoPort=35381, infoSecurePort=0, ipcPort=35747, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 1628869 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0xa4871f7444f09: from storage DS-ccdddced-bb98-4bc9-9df8-d7f561692c23 node DatanodeRegistration(127.0.0.1:40969, datanodeUuid=6fa86369-873f-4f5b-b41f-1e017be7aac9, infoPort=35381, infoSecurePort=0, ipcPort=35747, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 1628872 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985
   [junit4]   2> 1628873 WARN  (1512157112@qtp-148498253-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985) [    ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is false. Rechecking.
   [junit4]   2> 1628883 WARN  (1512157112@qtp-148498253-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985) [    ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is true
   [junit4]   2> 1628968 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data4/]]  heartbeating to localhost.localdomain/127.0.0.1:46029) [    ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
   [junit4]   2> 1628972 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0xa4871fd6f899b: from storage DS-ed7e664f-037a-4c90-9067-4d711de779af node DatanodeRegistration(127.0.0.1:45799, datanodeUuid=9a5392f1-a065-499d-a8b1-5fa36b32ee17, infoPort=37065, infoSecurePort=0, ipcPort=40531, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 1628973 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0xa4871fd6f899b: from storage DS-d1098035-70ce-418b-964e-b4e17ae4ead3 node DatanodeRegistration(127.0.0.1:45799, datanodeUuid=9a5392f1-a065-499d-a8b1-5fa36b32ee17, infoPort=37065, infoSecurePort=0, ipcPort=40531, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 1629029 INFO  (TEST-HdfsAutoAddReplicasIntegrationTest.testSimple-seed#[A483062C73879C9B]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testSimple
   [junit4]   2> 1629030 INFO  (qtp1940206296-28728) [n:127.0.0.1:34223_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf&maxShardsPerNode=2&autoAddReplicas=true&name=testSimple1&nrtReplicas=2&action=CREATE&numShards=2&createNodeSet=127.0.0.1:45039_solr,127.0.0.1:39879_solr&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1629032 INFO  (OverseerThreadFactory-8464-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection testSimple1
   [junit4]   2> 1629135 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"testSimple1",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"testSimple1_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:39879/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 1629136 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"testSimple1",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"testSimple1_shard1_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:45039/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 1629137 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"testSimple1",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"testSimple1_shard2_replica_n4",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:39879/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 1629138 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"testSimple1",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"testSimple1_shard2_replica_n6",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:45039/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 1629290 DEBUG (ScheduledTrigger-8462-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 3
   [junit4]   2> 1629340 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr    x:testSimple1_shard2_replica_n4] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf&newCollection=true&name=testSimple1_shard2_replica_n4&action=CREATE&numShards=2&collection=testSimple1&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1629340 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr    x:testSimple1_shard1_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf&newCollection=true&name=testSimple1_shard1_replica_n2&action=CREATE&numShards=2&collection=testSimple1&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1629341 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr    x:testSimple1_shard1_replica_n2] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1629341 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr    x:testSimple1_shard2_replica_n6] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf&newCollection=true&name=testSimple1_shard2_replica_n6&action=CREATE&numShards=2&collection=testSimple1&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1629343 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr    x:testSimple1_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf&newCollection=true&name=testSimple1_shard1_replica_n1&action=CREATE&numShards=2&collection=testSimple1&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1630291 DEBUG (ScheduledTrigger-8462-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 3
   [junit4]   2> 1630346 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1630346 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1630360 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1630362 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1630363 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.s.IndexSchema [testSimple1_shard2_replica_n4] Schema name=minimal
   [junit4]   2> 1630363 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.s.IndexSchema [testSimple1_shard2_replica_n6] Schema name=minimal
   [junit4]   2> 1630365 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 1630365 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard2_replica_n4' using configuration from collection testSimple1, trusted=true
   [junit4]   2> 1630366 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 1630366 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard2_replica_n6' using configuration from collection testSimple1, trusted=true
   [junit4]   2> 1630366 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard2.replica_n4' (registry 'solr.core.testSimple1.shard2.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1630366 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard2.replica_n6' (registry 'solr.core.testSimple1.shard2.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1630367 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.s.IndexSchema [testSimple1_shard1_replica_n2] Schema name=minimal
   [junit4]   2> 1630378 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.s.IndexSchema [testSimple1_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 1630378 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 1630379 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard1_replica_n2' using configuration from collection testSimple1, trusted=true
   [junit4]   2> 1630379 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard1.replica_n2' (registry 'solr.core.testSimple1.shard1.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1630379 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 1630379 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard1_replica_n1' using configuration from collection testSimple1, trusted=true
   [junit4]   2> 1630383 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard1.replica_n1' (registry 'solr.core.testSimple1.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1630385 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory solr.hdf

[...truncated too long message...]

build/contrib/solr-dataimporthandler/test/temp/junit4-J0-20180623_003429_74715312066844850799510.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-dataimporthandler/test/temp/junit4-J1-20180623_003429_7479827180275547199107.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 567 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J1-20180623_003457_7593588562413959039049.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J0-20180623_003457_7592461396733423052156.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J2-20180623_003457_76214734783176774942234.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 576 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J2-20180623_003510_19510980844002221148259.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J0-20180623_003510_1958195709446057798390.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J1-20180623_003510_1952916833743249538977.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 674 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J2-20180623_003520_44115365425148263749297.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J0-20180623_003520_44116354916187501290760.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J1-20180623_003520_44113703101826179663573.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 570 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-prometheus-exporter/test/temp/junit4-J0-20180623_003550_60611854339242176193155.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-prometheus-exporter/test/temp/junit4-J1-20180623_003550_6069919291497190540503.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 606 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-uima/test/temp/junit4-J1-20180623_003611_4283766516359282848020.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-uima/test/temp/junit4-J0-20180623_003611_42815424760474818802855.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 557 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-velocity/test/temp/junit4-J0-20180623_003625_2359907790589491311819.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 34353 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22300/consoleText

[repro] Revision: 3b9d3a760a432b97aad2c08b2f778fa2344eb14a

[repro] Ant options: "-Dargs=-XX:-UseCompressedOops -XX:+UseConcMarkSweepGC"
[repro] Repro line:  ant test  -Dtestcase=HdfsAutoAddReplicasIntegrationTest -Dtests.method=testSimple -Dtests.seed=A483062C73879C9B -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ga -Dtests.timezone=Europe/Amsterdam -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1

[repro] ant clean

[...truncated 6 lines...]
[repro] Test suites by module:
[repro]    solr/core
[repro]       HdfsAutoAddReplicasIntegrationTest
[repro] ant compile-test

[...truncated 3305 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.HdfsAutoAddReplicasIntegrationTest" -Dtests.showOutput=onerror "-Dargs=-XX:-UseCompressedOops -XX:+UseConcMarkSweepGC" -Dtests.seed=A483062C73879C9B -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ga -Dtests.timezone=Europe/Amsterdam -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1

[...truncated 45 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J0-20180623_003921_9778145130587091982603.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 4 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J2-20180623_003921_97710617184633399344356.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 4 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J1-20180623_003921_97713048276681065257358.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 23 lines...]
[repro] Failures:
[repro]   0/5 failed: org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest
[repro] Exiting with code 0

[...truncated 40 lines...]

[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-10) - Build # 22299 - Still Unstable!

Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22299/
Java: 64bit/jdk-10 -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC

2 tests failed.
FAILED:  org.apache.solr.cloud.cdcr.CdcrBidirectionalTest.testBiDir

Error Message:
Captured an uncaught exception in thread: Thread[id=1434, name=cdcr-replicator-641-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]

Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=1434, name=cdcr-replicator-641-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
Caused by: java.lang.AssertionError
	at __randomizedtesting.SeedInfo.seed([179CC4BE3CA5082E]:0)
	at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
	at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
	at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
	at java.base/java.lang.Thread.run(Thread.java:844)


FAILED:  org.apache.solr.schema.TestBulkSchemaConcurrent.test

Error Message:
Captured an uncaught exception in thread: Thread[id=34281, name=Thread-7131, state=RUNNABLE, group=TGRP-TestBulkSchemaConcurrent]

Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=34281, name=Thread-7131, state=RUNNABLE, group=TGRP-TestBulkSchemaConcurrent]
	at __randomizedtesting.SeedInfo.seed([179CC4BE3CA5082E:9FC8FB64925965D6]:0)
Caused by: java.lang.ArrayIndexOutOfBoundsException: 1
	at __randomizedtesting.SeedInfo.seed([179CC4BE3CA5082E]:0)
	at java.base/java.util.ArrayList.add(ArrayList.java:468)
	at java.base/java.util.ArrayList.add(ArrayList.java:480)
	at org.apache.solr.schema.TestBulkSchemaConcurrent$1.run(TestBulkSchemaConcurrent.java:71)




Build Log:
[...truncated 1869 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J1-20180622_193122_97517226885698389146925.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J2-20180622_193122_9757075924846183082882.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J0-20180622_193122_975442950508919956658.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 296 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J2-20180622_193952_3193686486944885433337.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J0-20180622_193952_3192405921133633563316.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J1-20180622_193952_33712965927917989661025.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 1080 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J0-20180622_194142_7623912414853969005435.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J2-20180622_194142_7629445416046634911711.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J1-20180622_194142_76216092446277168437633.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 258 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J2-20180622_194335_5905188593965878184353.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J0-20180622_194335_5905907319949510649270.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J1-20180622_194335_59016934120050157469507.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 253 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J0-20180622_194352_1891524542333264745157.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J2-20180622_194352_18918327593702397342560.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J1-20180622_194352_18914295979764365092854.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 162 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J1-20180622_194416_50511010918296710926313.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J2-20180622_194416_5057892806941758000679.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J0-20180622_194416_50512141276080296475825.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 204 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J0-20180622_194420_8444738997090051846087.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J2-20180622_194420_844557805422627418046.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J1-20180622_194420_844999137335804839095.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 171 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J1-20180622_194431_1593528630143478597930.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J0-20180622_194431_1593017240171163191103.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J2-20180622_194431_1597358790664321600406.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 176 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J0-20180622_194434_21911112959124792729672.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J1-20180622_194434_21915846851019055467493.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J2-20180622_194434_2199245096596122481454.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 161 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/smartcn/test/temp/junit4-J0-20180622_194449_94813458222355665369086.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/smartcn/test/temp/junit4-J1-20180622_194449_94813400652518548487429.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 162 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J2-20180622_194458_35916276298283876423441.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J1-20180622_194458_35913824030595924500105.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J0-20180622_194458_35918394980111726179542.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 177 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/uima/test/temp/junit4-J0-20180622_194501_03010756028188481934894.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 158 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J1-20180622_194510_0543407066472000457286.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J2-20180622_194510_05415380270914633833292.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 19 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J0-20180622_194510_053526140071587305912.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1391 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J2-20180622_194536_5233051038126053632763.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J1-20180622_194536_52310144032591434606380.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J0-20180622_194536_5234479507812897990374.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 250 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J0-20180622_194545_3208419766600881141627.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J1-20180622_194545_32016769179958799114602.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J2-20180622_194545_32016686905463552599960.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 256 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J2-20180622_194556_9887266072389824694675.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J1-20180622_194556_98814383853960319890556.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J0-20180622_194556_9883309568283475684067.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 228 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J1-20180622_194849_5807838831419435511447.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J2-20180622_194849_58018195436423336817804.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J0-20180622_194849_58016108983561728684701.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 176 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J0-20180622_194853_30815284055458188075972.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J1-20180622_194853_30814882213002732260739.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J2-20180622_194853_3082593387241268239698.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 235 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J1-20180622_194859_7223631303731788626096.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J0-20180622_194859_72216631581888688745214.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J2-20180622_194859_7228318924667308490901.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 185 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J2-20180622_194935_9248044510947296136467.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J1-20180622_194935_92415790082328535306166.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J0-20180622_194935_92410436223115756828841.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 249 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J2-20180622_194947_7462705089181124680576.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 12 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J0-20180622_194947_74618008694241034078189.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J1-20180622_194947_7463404260781484468196.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 163 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J2-20180622_195011_79385458384808107607.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J0-20180622_195011_79310374702511671455029.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J1-20180622_195011_7932396952698479916059.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 155 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/memory/test/temp/junit4-J0-20180622_195039_70017585345462959153639.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/memory/test/temp/junit4-J1-20180622_195039_70013203778464665042643.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 187 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J1-20180622_195056_8189925253060255085402.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J2-20180622_195056_81813245576896810006276.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J0-20180622_195056_81817572480428787338831.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 307 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J2-20180622_195114_5256618398424970030131.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J0-20180622_195114_5252374566141527664182.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J1-20180622_195114_5255781932052843908505.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 229 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J0-20180622_195123_44013261080192548887106.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J2-20180622_195123_4459940662138801853729.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J1-20180622_195123_4459290371273100067998.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 204 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J1-20180622_195132_74417327947547836054515.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J0-20180622_195132_74414840037538401637989.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 8 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J2-20180622_195132_7446766959302487954346.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 229 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J2-20180622_195144_9389598586621052203095.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J1-20180622_195144_93810165008891834608894.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J0-20180622_195144_93815598487795546639998.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 298 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J0-20180622_195256_63613461926679787916503.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J2-20180622_195256_63617194451937851710965.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J1-20180622_195256_6361755680806213457657.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 153 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial/test/temp/junit4-J0-20180622_195334_04010767203489090143056.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 181 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J2-20180622_195336_03111806640830895513871.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J0-20180622_195336_0314992987153892666295.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 11 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J1-20180622_195336_03117321978236797639185.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 252 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J0-20180622_195441_37413644268683172775391.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J2-20180622_195441_38012166670838900255214.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 9 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J1-20180622_195441_3769956743601339604812.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 1837 lines...]
   [junit4] Suite: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest
   [junit4]   2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/init-core-data-001
   [junit4]   2> 65879 WARN  (SUITE-CdcrBidirectionalTest-seed#[179CC4BE3CA5082E]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=21 numCloses=21
   [junit4]   2> 65879 INFO  (SUITE-CdcrBidirectionalTest-seed#[179CC4BE3CA5082E]-worker) [    ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 65880 INFO  (SUITE-CdcrBidirectionalTest-seed#[179CC4BE3CA5082E]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason="", ssl=0.0/0.0, value=0.0/0.0, clientAuth=0.0/0.0)
   [junit4]   2> 65880 INFO  (SUITE-CdcrBidirectionalTest-seed#[179CC4BE3CA5082E]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 65882 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testBiDir
   [junit4]   2> 65884 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 1 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster2-001
   [junit4]   2> 65884 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 65885 INFO  (Thread-303) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 65885 INFO  (Thread-303) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 65930 ERROR (Thread-303) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 65985 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.ZkTestServer start zk server on port:43011
   [junit4]   2> 65987 INFO  (zkConnectionManagerCallback-274-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66003 INFO  (jetty-launcher-271-thread-1) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
   [junit4]   2> 66039 INFO  (jetty-launcher-271-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 66039 INFO  (jetty-launcher-271-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 66039 INFO  (jetty-launcher-271-thread-1) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 66040 INFO  (jetty-launcher-271-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@1318fb9e{/solr,null,AVAILABLE}
   [junit4]   2> 66041 INFO  (jetty-launcher-271-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@2a29e217{SSL,[ssl, http/1.1]}{127.0.0.1:34323}
   [junit4]   2> 66041 INFO  (jetty-launcher-271-thread-1) [    ] o.e.j.s.Server Started @66070ms
   [junit4]   2> 66041 INFO  (jetty-launcher-271-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=34323}
   [junit4]   2> 66042 ERROR (jetty-launcher-271-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 66042 INFO  (jetty-launcher-271-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 66042 INFO  (jetty-launcher-271-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 8.0.0
   [junit4]   2> 66042 INFO  (jetty-launcher-271-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 66042 INFO  (jetty-launcher-271-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 66042 INFO  (jetty-launcher-271-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-22T19:57:10.052430Z
   [junit4]   2> 66050 INFO  (zkConnectionManagerCallback-276-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66050 INFO  (jetty-launcher-271-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 66173 INFO  (jetty-launcher-271-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:43011/solr
   [junit4]   2> 66174 INFO  (zkConnectionManagerCallback-280-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66175 INFO  (zkConnectionManagerCallback-282-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66248 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 66248 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:34323_solr
   [junit4]   2> 66249 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.c.Overseer Overseer (id=72246242139963395-127.0.0.1:34323_solr-n_0000000000) starting
   [junit4]   2> 66253 INFO  (zkConnectionManagerCallback-289-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66255 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:43011/solr ready
   [junit4]   2> 66278 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34323_solr
   [junit4]   2> 66300 INFO  (zkCallback-281-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 66328 INFO  (zkCallback-288-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 66373 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 66392 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34323.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 66409 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34323.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 66409 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34323.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 66410 INFO  (jetty-launcher-271-thread-1) [n:127.0.0.1:34323_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster2-001/node1/.
   [junit4]   2> 66517 INFO  (zkConnectionManagerCallback-292-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66597 INFO  (zkConnectionManagerCallback-295-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66599 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 1 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster1-001
   [junit4]   2> 66599 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 66605 INFO  (Thread-313) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 66606 INFO  (Thread-313) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 66646 ERROR (Thread-313) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 66705 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.ZkTestServer start zk server on port:43693
   [junit4]   2> 66750 INFO  (zkConnectionManagerCallback-299-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66768 INFO  (jetty-launcher-296-thread-1) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
   [junit4]   2> 66807 INFO  (jetty-launcher-296-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 66807 INFO  (jetty-launcher-296-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 66807 INFO  (jetty-launcher-296-thread-1) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 66810 INFO  (jetty-launcher-296-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@6e618b3e{/solr,null,AVAILABLE}
   [junit4]   2> 66810 INFO  (jetty-launcher-296-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@10be4e6a{SSL,[ssl, http/1.1]}{127.0.0.1:38339}
   [junit4]   2> 66810 INFO  (jetty-launcher-296-thread-1) [    ] o.e.j.s.Server Started @66839ms
   [junit4]   2> 66810 INFO  (jetty-launcher-296-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=38339}
   [junit4]   2> 66811 ERROR (jetty-launcher-296-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 66811 INFO  (jetty-launcher-296-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 66811 INFO  (jetty-launcher-296-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 8.0.0
   [junit4]   2> 66811 INFO  (jetty-launcher-296-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 66811 INFO  (jetty-launcher-296-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 66811 INFO  (jetty-launcher-296-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-22T19:57:10.821676Z
   [junit4]   2> 66818 INFO  (zkConnectionManagerCallback-301-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 66819 INFO  (jetty-launcher-296-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 66944 INFO  (jetty-launcher-296-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:43693/solr
   [junit4]   2> 66973 INFO  (zkConnectionManagerCallback-305-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 67005 INFO  (zkConnectionManagerCallback-307-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 67093 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 67094 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:38339_solr
   [junit4]   2> 67094 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.c.Overseer Overseer (id=72246242186952707-127.0.0.1:38339_solr-n_0000000000) starting
   [junit4]   2> 67115 INFO  (zkConnectionManagerCallback-314-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 67116 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:43693/solr ready
   [junit4]   2> 67128 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:38339_solr
   [junit4]   2> 67155 INFO  (zkCallback-306-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 67155 INFO  (zkCallback-313-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 67177 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 67211 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38339.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 67222 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38339.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 67222 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38339.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 67223 INFO  (jetty-launcher-296-thread-1) [n:127.0.0.1:38339_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster1-001/node1/.
   [junit4]   2> 67270 INFO  (zkConnectionManagerCallback-317-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 67294 INFO  (zkConnectionManagerCallback-320-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 67295 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.c.CdcrBidirectionalTest cluster2 zkHost = 127.0.0.1:43011/solr
   [junit4]   2> 67295 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.c.CdcrBidirectionalTest cluster1 zkHost = 127.0.0.1:43693/solr
   [junit4]   2> 67304 INFO  (zkConnectionManagerCallback-322-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 67324 INFO  (zkConnectionManagerCallback-326-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 67325 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 67325 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:43693/solr ready
   [junit4]   2> 67352 INFO  (qtp1271965624-1289) [n:127.0.0.1:38339_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=cdcr-cluster1&maxShardsPerNode=2&name=cdcr-cluster1&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 67358 INFO  (OverseerThreadFactory-602-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection cdcr-cluster1
   [junit4]   2> 67491 INFO  (OverseerStateUpdate-72246242186952707-127.0.0.1:38339_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster1",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"cdcr-cluster1_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:38339/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 67492 INFO  (OverseerStateUpdate-72246242186952707-127.0.0.1:38339_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster1",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"cdcr-cluster1_shard2_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:38339/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 67776 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr    x:cdcr-cluster1_shard2_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster1_shard2_replica_n2&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 67776 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr    x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node3&name=cdcr-cluster1_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 67776 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr    x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 68817 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 68817 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 68831 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.s.IndexSchema [cdcr-cluster1_shard2_replica_n2] Schema name=minimal
   [junit4]   2> 68832 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 68832 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster1_shard2_replica_n2' using configuration from collection cdcr-cluster1, trusted=true
   [junit4]   2> 68838 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38339.solr.core.cdcr-cluster1.shard2.replica_n2' (registry 'solr.core.cdcr-cluster1.shard2.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 68841 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 68841 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.SolrCore [[cdcr-cluster1_shard2_replica_n2] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster1-001/node1/cdcr-cluster1_shard2_replica_n2], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster1-001/node1/./cdcr-cluster1_shard2_replica_n2/data/]
   [junit4]   2> 68846 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.IndexSchema [cdcr-cluster1_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 68847 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 68847 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster1_shard1_replica_n1' using configuration from collection cdcr-cluster1, trusted=true
   [junit4]   2> 68848 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_38339.solr.core.cdcr-cluster1.shard1.replica_n1' (registry 'solr.core.cdcr-cluster1.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 68848 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 68848 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [[cdcr-cluster1_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster1-001/node1/cdcr-cluster1_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster1-001/node1/./cdcr-cluster1_shard1_replica_n1/data/]
   [junit4]   2> 68930 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 68941 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 68936 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 68941 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 68942 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 68943 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 68943 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 68946 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@6a463c32[cdcr-cluster1_shard1_replica_n1] main]
   [junit4]   2> 68947 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster1
   [junit4]   2> 68948 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster1
   [junit4]   2> 68950 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster1/managed-schema
   [junit4]   2> 68951 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 68951 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 68955 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrBufferStateManager Created znode /collections/cdcr-cluster1/cdcr/state/buffer
   [junit4]   2> 68942 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 68957 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrProcessStateManager Created znode /collections/cdcr-cluster1/cdcr/state/process
   [junit4]   2> 68960 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@538716fb[cdcr-cluster1_shard2_replica_n2] main]
   [junit4]   2> 68961 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster1
   [junit4]   2> 68961 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster1
   [junit4]   2> 68962 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster1/managed-schema
   [junit4]   2> 68962 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 68963 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 68974 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604004015488630784
   [junit4]   2> 68975 INFO  (searcherExecutor-608-thread-1-processing-n:127.0.0.1:38339_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node3) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@6a463c32[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 68975 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604004015489679360
   [junit4]   2> 68978 INFO  (searcherExecutor-607-thread-1-processing-n:127.0.0.1:38339_solr x:cdcr-cluster1_shard2_replica_n2 c:cdcr-cluster1 s:shard2 r:core_node4) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.SolrCore [cdcr-cluster1_shard2_replica_n2] Registered new searcher Searcher@538716fb[cdcr-cluster1_shard2_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 68979 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard1 to Terms{values={core_node3=0}, version=0}
   [junit4]   2> 68980 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard2 to Terms{values={core_node4=0}, version=0}
   [junit4]   2> 68981 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 68981 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 68981 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:38339/solr/cdcr-cluster1_shard1_replica_n1/
   [junit4]   2> 68981 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 68981 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:38339/solr/cdcr-cluster1_shard1_replica_n1/ has no replicas
   [junit4]   2> 68981 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 68982 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 68982 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 68982 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:38339/solr/cdcr-cluster1_shard2_replica_n2/
   [junit4]   2> 68982 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 68982 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.SyncStrategy https://127.0.0.1:38339/solr/cdcr-cluster1_shard2_replica_n2/ has no replicas
   [junit4]   2> 68982 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 68990 INFO  (zkCallback-306-thread-3) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster1:shard1
   [junit4]   2> 68991 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:38339/solr/cdcr-cluster1_shard1_replica_n1/ shard1
   [junit4]   2> 68992 INFO  (zkCallback-306-thread-3) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster1:shard2
   [junit4]   2> 68998 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:38339/solr/cdcr-cluster1_shard2_replica_n2/ shard2
   [junit4]   2> 69105 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 69112 INFO  (qtp1271965624-1290) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n2] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster1_shard2_replica_n2&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1336
   [junit4]   2> 69144 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 69145 INFO  (qtp1271965624-1287) [n:127.0.0.1:38339_solr c:cdcr-cluster1 s:shard1 r:core_node3 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node3&name=cdcr-cluster1_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1369
   [junit4]   2> 69148 INFO  (qtp1271965624-1289) [n:127.0.0.1:38339_solr    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 69246 INFO  (zkCallback-306-thread-3) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/state.json] for collection [cdcr-cluster1] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 69360 INFO  (OverseerCollectionConfigSetProcessor-72246242186952707-127.0.0.1:38339_solr-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 70150 INFO  (qtp1271965624-1289) [n:127.0.0.1:38339_solr    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=cdcr-cluster1&maxShardsPerNode=2&name=cdcr-cluster1&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2} status=0 QTime=2798
   [junit4]   2> 70161 INFO  (zkConnectionManagerCallback-330-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 70173 INFO  (zkConnectionManagerCallback-334-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 70178 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 70178 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[179CC4BE3CA5082E]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:43011/solr ready
   [junit4]   2> 70215 INFO  (qtp894101829-1235) [n:127.0.0.1:34323_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=cdcr-cluster2&maxShardsPerNode=2&name=cdcr-cluster2&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 70225 INFO  (OverseerThreadFactory-589-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection cdcr-cluster2
   [junit4]   2> 70334 INFO  (OverseerStateUpdate-72246242139963395-127.0.0.1:34323_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster2",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"cdcr-cluster2_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:34323/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 70337 INFO  (OverseerStateUpdate-72246242139963395-127.0.0.1:34323_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster2",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"cdcr-cluster2_shard2_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:34323/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 70548 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr    x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node3&name=cdcr-cluster2_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 70548 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr    x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 70549 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr    x:cdcr-cluster2_shard2_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster2_shard2_replica_n2&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 71557 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 71561 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.IndexSchema [cdcr-cluster2_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 71562 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 71562 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster2_shard1_replica_n1' using configuration from collection cdcr-cluster2, trusted=true
   [junit4]   2> 71563 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34323.solr.core.cdcr-cluster2.shard1.replica_n1' (registry 'solr.core.cdcr-cluster2.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 71563 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 71563 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [[cdcr-cluster2_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster2-001/node1/cdcr-cluster2_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster2-001/node1/./cdcr-cluster2_shard1_replica_n1/data/]
   [junit4]   2> 71572 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 71577 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.s.IndexSchema [cdcr-cluster2_shard2_replica_n2] Schema name=minimal
   [junit4]   2> 71578 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 71578 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster2_shard2_replica_n2' using configuration from collection cdcr-cluster2, trusted=true
   [junit4]   2> 71578 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34323.solr.core.cdcr-cluster2.shard2.replica_n2' (registry 'solr.core.cdcr-cluster2.shard2.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@1bdae3c2
   [junit4]   2> 71578 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 71578 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.c.SolrCore [[cdcr-cluster2_shard2_replica_n2] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster2-001/node1/cdcr-cluster2_shard2_replica_n2], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_179CC4BE3CA5082E-001/cdcr-cluster2-001/node1/./cdcr-cluster2_shard2_replica_n2/data/]
   [junit4]   2> 71679 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 71679 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 71680 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 71680 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 71683 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@66532b2d[cdcr-cluster2_shard1_replica_n1] main]
   [junit4]   2> 71684 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 71684 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster2
   [junit4]   2> 71684 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 71685 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 71685 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 71686 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster2
   [junit4]   2> 71686 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster2/managed-schema
   [junit4]   2> 71686 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 71687 INFO  (qtp894101829-1237) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard1 r:core_node3 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 71688 INFO  (qtp894101829-1368) [n:127.0.0.1:34323_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@10e6f36[cdcr-cluster2_sha

[...truncated too long message...]

 J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-dataimporthandler/test/temp/junit4-J2-20180622_203347_36612163132345136366026.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 567 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J0-20180622_203409_7312533074859934614193.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J2-20180622_203409_7312100833796547812976.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J1-20180622_203409_73110200560447635004858.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 573 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J1-20180622_203420_8495781580330671065330.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J0-20180622_203420_8498127815912000868131.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J2-20180622_203420_84912412993694477396474.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 674 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J1-20180622_203430_5118442383444320962715.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J2-20180622_203430_5111186050719970513248.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J0-20180622_203430_511998131075285183705.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 570 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-prometheus-exporter/test/temp/junit4-J1-20180622_203500_4723401477613133505857.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-prometheus-exporter/test/temp/junit4-J0-20180622_203500_4729805252188485125067.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 606 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-uima/test/temp/junit4-J1-20180622_203520_04716979633493213653943.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-uima/test/temp/junit4-J0-20180622_203520_04712010436327617011160.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 557 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-velocity/test/temp/junit4-J0-20180622_203532_2882621972915450260621.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 34353 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22299/consoleText

[repro] Revision: 3b9d3a760a432b97aad2c08b2f778fa2344eb14a

[repro] Ant options: "-Dargs=-XX:+UseCompressedOops -XX:+UseConcMarkSweepGC"
[repro] Repro line:  ant test  -Dtestcase=CdcrBidirectionalTest -Dtests.method=testBiDir -Dtests.seed=179CC4BE3CA5082E -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=az-Latn -Dtests.timezone=Australia/Queensland -Dtests.asserts=true -Dtests.file.encoding=US-ASCII

[repro] Repro line:  ant test  -Dtestcase=TestBulkSchemaConcurrent -Dtests.method=test -Dtests.seed=179CC4BE3CA5082E -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=en-AS -Dtests.timezone=Africa/Gaborone -Dtests.asserts=true -Dtests.file.encoding=US-ASCII

[repro] ant clean

[...truncated 6 lines...]
[repro] Test suites by module:
[repro]    solr/core
[repro]       TestBulkSchemaConcurrent
[repro]       CdcrBidirectionalTest
[repro] ant compile-test

[...truncated 3305 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=10 -Dtests.class="*.TestBulkSchemaConcurrent|*.CdcrBidirectionalTest" -Dtests.showOutput=onerror "-Dargs=-XX:+UseCompressedOops -XX:+UseConcMarkSweepGC" -Dtests.seed=179CC4BE3CA5082E -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=en-AS -Dtests.timezone=Africa/Gaborone -Dtests.asserts=true -Dtests.file.encoding=US-ASCII

[...truncated 66 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J1-20180622_203822_84410286526031136525859.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 4 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J2-20180622_203822_8447287561494765356348.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 4 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J0-20180622_203822_844592174227435754310.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 23 lines...]
[repro] Failures:
[repro]   0/5 failed: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest
[repro]   0/5 failed: org.apache.solr.schema.TestBulkSchemaConcurrent
[repro] Exiting with code 0

[...truncated 40 lines...]