You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2018/06/23 00:45:26 UTC

[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-10) - Build # 22300 - Still Unstable!

Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22300/
Java: 64bit/jdk-10 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest.testSimple

Error Message:
Waiting for collection testSimple2 null Live Nodes: [127.0.0.1:34223_solr, 127.0.0.1:45039_solr] Last available state: DocCollection(testSimple2//collections/testSimple2/state.json/21)={   "pullReplicas":"0",   "replicationFactor":"2",   "shards":{     "shard1":{       "range":"80000000-ffffffff",       "state":"active",       "replicas":{         "core_node3":{           "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/",           "base_url":"http://127.0.0.1:34223/solr",           "node_name":"127.0.0.1:34223_solr",           "type":"NRT",           "force_set_state":"false",           "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/tlog",           "core":"testSimple2_shard1_replica_n1",           "shared_storage":"true",           "state":"active",           "leader":"true"},         "core_node5":{           "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/",           "base_url":"http://127.0.0.1:39879/solr",           "node_name":"127.0.0.1:39879_solr",           "type":"NRT",           "force_set_state":"false",           "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/tlog",           "core":"testSimple2_shard1_replica_n2",           "shared_storage":"true",           "state":"down"}}},     "shard2":{       "range":"0-7fffffff",       "state":"active",       "replicas":{         "core_node6":{           "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/",           "base_url":"http://127.0.0.1:34223/solr",           "node_name":"127.0.0.1:34223_solr",           "type":"NRT",           "force_set_state":"false",           "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/tlog",           "core":"testSimple2_shard2_replica_n4",           "shared_storage":"true",           "state":"active",           "leader":"true"},         "core_node8":{           "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/",           "base_url":"http://127.0.0.1:39879/solr",           "node_name":"127.0.0.1:39879_solr",           "type":"NRT",           "force_set_state":"false",           "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/tlog",           "core":"testSimple2_shard2_replica_n7",           "shared_storage":"true",           "state":"down"}}}},   "router":{"name":"compositeId"},   "maxShardsPerNode":"2",   "autoAddReplicas":"true",   "nrtReplicas":"2",   "tlogReplicas":"0"}

Stack Trace:
java.lang.AssertionError: Waiting for collection testSimple2
null
Live Nodes: [127.0.0.1:34223_solr, 127.0.0.1:45039_solr]
Last available state: DocCollection(testSimple2//collections/testSimple2/state.json/21)={
  "pullReplicas":"0",
  "replicationFactor":"2",
  "shards":{
    "shard1":{
      "range":"80000000-ffffffff",
      "state":"active",
      "replicas":{
        "core_node3":{
          "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/",
          "base_url":"http://127.0.0.1:34223/solr",
          "node_name":"127.0.0.1:34223_solr",
          "type":"NRT",
          "force_set_state":"false",
          "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/tlog",
          "core":"testSimple2_shard1_replica_n1",
          "shared_storage":"true",
          "state":"active",
          "leader":"true"},
        "core_node5":{
          "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/",
          "base_url":"http://127.0.0.1:39879/solr",
          "node_name":"127.0.0.1:39879_solr",
          "type":"NRT",
          "force_set_state":"false",
          "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/tlog",
          "core":"testSimple2_shard1_replica_n2",
          "shared_storage":"true",
          "state":"down"}}},
    "shard2":{
      "range":"0-7fffffff",
      "state":"active",
      "replicas":{
        "core_node6":{
          "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/",
          "base_url":"http://127.0.0.1:34223/solr",
          "node_name":"127.0.0.1:34223_solr",
          "type":"NRT",
          "force_set_state":"false",
          "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/tlog",
          "core":"testSimple2_shard2_replica_n4",
          "shared_storage":"true",
          "state":"active",
          "leader":"true"},
        "core_node8":{
          "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/",
          "base_url":"http://127.0.0.1:39879/solr",
          "node_name":"127.0.0.1:39879_solr",
          "type":"NRT",
          "force_set_state":"false",
          "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/tlog",
          "core":"testSimple2_shard2_replica_n7",
          "shared_storage":"true",
          "state":"down"}}}},
  "router":{"name":"compositeId"},
  "maxShardsPerNode":"2",
  "autoAddReplicas":"true",
  "nrtReplicas":"2",
  "tlogReplicas":"0"}
	at __randomizedtesting.SeedInfo.seed([A483062C73879C9B:9C3022D25474484A]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.apache.solr.cloud.SolrCloudTestCase.waitForState(SolrCloudTestCase.java:278)
	at org.apache.solr.cloud.autoscaling.AutoAddReplicasIntegrationTest.testSimple(AutoAddReplicasIntegrationTest.java:129)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.base/java.lang.reflect.Method.invoke(Method.java:564)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 1871 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J0-20180622_233120_8438624310112548928347.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J2-20180622_233120_843109218256741183964.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J1-20180622_233120_84215878213975654336568.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 298 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J1-20180622_234003_82713161516520257880618.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J2-20180622_234003_82211985366498351429167.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J0-20180622_234003_8187629202302504117049.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1080 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J0-20180622_234142_9048161528471907446887.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J2-20180622_234142_90416258816079129641780.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J1-20180622_234142_90412005419182290644372.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 258 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J0-20180622_234332_4914664188095482509082.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J2-20180622_234332_4922443722992351897655.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J1-20180622_234332_4914364283643442390734.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 250 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J0-20180622_234345_5569241898986794068605.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J2-20180622_234345_55612839679531665619473.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J1-20180622_234345_55614100303459436628584.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 162 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J1-20180622_234408_1182183404860930597018.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J0-20180622_234408_11817221358498753266703.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J2-20180622_234408_1191547639190121044076.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 210 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J0-20180622_234412_40714371405619675624954.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J2-20180622_234412_4078848850315630926441.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J1-20180622_234412_4079975389867110683768.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 174 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J1-20180622_234421_51010637845793649127342.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J0-20180622_234421_5103569472874704098846.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J2-20180622_234421_51016436102187679443010.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 176 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J2-20180622_234424_31610752690720732553073.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J0-20180622_234424_3163920468419256368342.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J1-20180622_234424_31613507803586776163227.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 161 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/smartcn/test/temp/junit4-J0-20180622_234434_0177751930883938230811.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/smartcn/test/temp/junit4-J1-20180622_234434_01715792206888438466742.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 162 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J1-20180622_234443_3098926868024307710344.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J2-20180622_234443_30910390262105719088600.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J0-20180622_234443_3078828821911177442895.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 177 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/uima/test/temp/junit4-J0-20180622_234446_5954096128335448572284.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 153 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J1-20180622_234455_65114409003380652460182.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J2-20180622_234455_6514623326656775997862.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 19 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J0-20180622_234455_6517020648835771442894.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1391 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J2-20180622_234519_1932894334289275466381.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J1-20180622_234519_1932449911797630396956.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J0-20180622_234519_1936352665634460400351.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 253 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J2-20180622_234527_4918016004016563548263.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J0-20180622_234527_49118385300419579868577.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J1-20180622_234527_4915089559484994378250.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 262 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J1-20180622_234539_34511673298149119415382.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J0-20180622_234539_3442056690456979073956.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J2-20180622_234539_3456681254644361673058.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 231 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J1-20180622_234754_3966129908289707100345.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J2-20180622_234754_39610786966458196672386.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J0-20180622_234754_39611330166380200542034.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 176 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J1-20180622_234757_3282744005959643641695.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J2-20180622_234757_328184188593405154770.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J0-20180622_234757_32813663791879804413349.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 235 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J1-20180622_234803_6451920093429510786067.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J2-20180622_234803_6456139679212212410020.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J0-20180622_234803_64513176520531889858516.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 185 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J2-20180622_234842_66210931058644992560173.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J1-20180622_234842_66216254532774656442410.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J0-20180622_234842_6625491280472342315484.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 255 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J1-20180622_234900_48815603692318856114647.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J2-20180622_234900_48816107964350347376247.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J0-20180622_234900_48816853337107767940056.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 163 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J0-20180622_234916_9509003729064799529499.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J2-20180622_234916_9509833749328796722833.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J1-20180622_234916_9501938938754275132751.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 155 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/memory/test/temp/junit4-J0-20180622_234945_85111956510346757286560.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/memory/test/temp/junit4-J1-20180622_234945_8526925569142241457753.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 189 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J2-20180622_234957_89716020023611267993152.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J1-20180622_234957_89713374075461271625175.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J0-20180622_234957_897995069218042617454.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 304 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J1-20180622_235015_07816445064399041937754.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J2-20180622_235015_07817300623641250331463.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J0-20180622_235015_0777051079486333859241.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 229 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J2-20180622_235023_9761355228138200064238.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J0-20180622_235023_9765821768678419074163.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J1-20180622_235023_9764016929074376670969.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 212 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J1-20180622_235031_70213674486619839082375.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J0-20180622_235031_70211591443367436951168.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J2-20180622_235031_70217646439452202035793.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 227 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J0-20180622_235044_36518075386306318393681.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J2-20180622_235044_3655371863721534862338.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 7 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J1-20180622_235044_3653195148306886251101.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 298 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J0-20180622_235133_51311242690447766539259.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J2-20180622_235133_5131023607900679846491.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J1-20180622_235133_5137580074882706317880.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 153 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial/test/temp/junit4-J0-20180622_235158_4753421684692232253190.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 181 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J0-20180622_235200_23017643266495511437138.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J1-20180622_235200_2306779390704051548979.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J2-20180622_235200_23015255487659489383656.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 264 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J2-20180622_235320_44513781554556465979598.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J0-20180622_235320_4444365517420477492284.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J1-20180622_235320_4449138489014132900971.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3929 lines...]
   [junit4] Suite: org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest
   [junit4]   2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/init-core-data-001
   [junit4]   2> 1628046 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=1 numCloses=1
   [junit4]   2> 1628046 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.SolrTestCaseJ4 Using TrieFields (NUMERIC_POINTS_SYSPROP=false) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 1628047 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0)
   [junit4]   2> 1628047 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 1628047 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 3 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001
   [junit4]   2> 1628047 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 1628048 INFO  (Thread-4546) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 1628048 INFO  (Thread-4546) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 1628049 ERROR (Thread-4546) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 1628148 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.ZkTestServer start zk server on port:36131
   [junit4]   2> 1628149 INFO  (zkConnectionManagerCallback-10100-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1628151 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@4ebf9fd8{/solr,null,AVAILABLE}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@636e248d{/solr,null,AVAILABLE}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@3a13410e{/solr,null,AVAILABLE}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.AbstractConnector Started ServerConnector@295b2198{HTTP/1.1,[http/1.1]}{127.0.0.1:45039}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@dd199bf{HTTP/1.1,[http/1.1]}{127.0.0.1:39879}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.e.j.s.Server Started @1628177ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.AbstractConnector Started ServerConnector@7ef44222{HTTP/1.1,[http/1.1]}{127.0.0.1:34223}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.e.j.s.Server Started @1628177ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.e.j.s.Server Started @1628177ms
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=45039}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=39879}
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=34223}
   [junit4]   2> 1628152 ERROR (jetty-launcher-10097-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1628152 ERROR (jetty-launcher-10097-thread-3) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1628152 ERROR (jetty-launcher-10097-thread-2) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 8.0.0
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 8.0.0
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 8.0.0
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1628152 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T00:21:31.684042Z
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T00:21:31.684074Z
   [junit4]   2> 1628153 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T00:21:31.684084Z
   [junit4]   2> 1628153 INFO  (zkConnectionManagerCallback-10104-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628153 INFO  (zkConnectionManagerCallback-10106-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628153 INFO  (zkConnectionManagerCallback-10103-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628154 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 1628154 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 1628154 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1628155 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1628156 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1628156 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1628157 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1628243 INFO  (jetty-launcher-10097-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
   [junit4]   2> 1628244 INFO  (zkConnectionManagerCallback-10110-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628245 INFO  (zkConnectionManagerCallback-10112-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628251 INFO  (jetty-launcher-10097-thread-3) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
   [junit4]   2> 1628260 INFO  (zkConnectionManagerCallback-10118-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628265 INFO  (zkConnectionManagerCallback-10120-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628269 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1628269 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:45039_solr
   [junit4]   2> 1628270 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.Overseer Overseer (id=72247281650171911-127.0.0.1:45039_solr-n_0000000000) starting
   [junit4]   2> 1628281 INFO  (zkConnectionManagerCallback-10127-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628282 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
   [junit4]   2> 1628284 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:45039_solr
   [junit4]   2> 1628285 INFO  (zkCallback-10126-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1628285 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Adding .auto_add_replicas and .scheduled_maintenance triggers
   [junit4]   2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 1
   [junit4]   2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 1, lastZnodeVersion -1
   [junit4]   2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 1
   [junit4]   2> 1628289 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:45039_solr]
   [junit4]   2> 1628289 DEBUG (ScheduledTrigger-8462-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 1628292 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1628296 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 1628296 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 1, lastZnodeVersion 1
   [junit4]   2> 1628297 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1628297 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1628298 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1628298 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:39879_solr
   [junit4]   2> 1628299 INFO  (zkCallback-10119-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1628299 INFO  (zkCallback-10126-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1628305 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1628313 INFO  (zkCallback-10111-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1628317 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628317 INFO  (zkConnectionManagerCallback-10133-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628318 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 1628318 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
   [junit4]   2> 1628318 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1628323 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628323 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628324 INFO  (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node3/.
   [junit4]   2> 1628330 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628336 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628336 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628337 INFO  (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node1/.
   [junit4]   2> 1628408 INFO  (jetty-launcher-10097-thread-2) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
   [junit4]   2> 1628409 INFO  (zkConnectionManagerCallback-10138-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628410 INFO  (zkConnectionManagerCallback-10140-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628413 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 1628414 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 1628415 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1628415 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34223_solr
   [junit4]   2> 1628415 INFO  (zkCallback-10119-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628415 INFO  (zkCallback-10126-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628415 INFO  (zkCallback-10111-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628415 INFO  (zkCallback-10139-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628415 INFO  (zkCallback-10132-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 1628424 INFO  (zkConnectionManagerCallback-10147-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628425 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 1628425 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
   [junit4]   2> 1628426 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1628435 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628440 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628440 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1628441 INFO  (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node2/.
   [junit4]   2> 1628457 INFO  (zkConnectionManagerCallback-10150-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628459 INFO  (zkConnectionManagerCallback-10155-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1628459 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 1628460 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 1628518 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 1628524 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 1628525 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 1628531 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_localdomain_33425_hdfs____.4bsgsv/webapp
   [junit4]   2> 1628601 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:33425
   [junit4]   2> 1628678 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 1628679 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 1628683 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_34005_datanode____3r5l3v/webapp
   [junit4]   2> 1628761 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:34005
   [junit4]   2> 1628785 WARN  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 1628786 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 1628789 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_32985_datanode____.yy7xwi/webapp
   [junit4]   2> 1628864 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data2/]]  heartbeating to localhost.localdomain/127.0.0.1:46029) [    ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
   [junit4]   2> 1628869 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0xa4871f7444f09: from storage DS-53fc62e1-2ec1-4aa4-a0a1-da327bd2c668 node DatanodeRegistration(127.0.0.1:40969, datanodeUuid=6fa86369-873f-4f5b-b41f-1e017be7aac9, infoPort=35381, infoSecurePort=0, ipcPort=35747, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 1628869 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0xa4871f7444f09: from storage DS-ccdddced-bb98-4bc9-9df8-d7f561692c23 node DatanodeRegistration(127.0.0.1:40969, datanodeUuid=6fa86369-873f-4f5b-b41f-1e017be7aac9, infoPort=35381, infoSecurePort=0, ipcPort=35747, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 1628872 INFO  (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985
   [junit4]   2> 1628873 WARN  (1512157112@qtp-148498253-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985) [    ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is false. Rechecking.
   [junit4]   2> 1628883 WARN  (1512157112@qtp-148498253-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985) [    ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is true
   [junit4]   2> 1628968 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data4/]]  heartbeating to localhost.localdomain/127.0.0.1:46029) [    ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
   [junit4]   2> 1628972 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0xa4871fd6f899b: from storage DS-ed7e664f-037a-4c90-9067-4d711de779af node DatanodeRegistration(127.0.0.1:45799, datanodeUuid=9a5392f1-a065-499d-a8b1-5fa36b32ee17, infoPort=37065, infoSecurePort=0, ipcPort=40531, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 1628973 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0xa4871fd6f899b: from storage DS-d1098035-70ce-418b-964e-b4e17ae4ead3 node DatanodeRegistration(127.0.0.1:45799, datanodeUuid=9a5392f1-a065-499d-a8b1-5fa36b32ee17, infoPort=37065, infoSecurePort=0, ipcPort=40531, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 1629029 INFO  (TEST-HdfsAutoAddReplicasIntegrationTest.testSimple-seed#[A483062C73879C9B]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testSimple
   [junit4]   2> 1629030 INFO  (qtp1940206296-28728) [n:127.0.0.1:34223_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf&maxShardsPerNode=2&autoAddReplicas=true&name=testSimple1&nrtReplicas=2&action=CREATE&numShards=2&createNodeSet=127.0.0.1:45039_solr,127.0.0.1:39879_solr&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1629032 INFO  (OverseerThreadFactory-8464-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection testSimple1
   [junit4]   2> 1629135 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"testSimple1",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"testSimple1_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:39879/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 1629136 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"testSimple1",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"testSimple1_shard1_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:45039/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 1629137 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"testSimple1",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"testSimple1_shard2_replica_n4",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:39879/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 1629138 INFO  (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"testSimple1",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"testSimple1_shard2_replica_n6",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:45039/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 1629290 DEBUG (ScheduledTrigger-8462-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 3
   [junit4]   2> 1629340 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr    x:testSimple1_shard2_replica_n4] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf&newCollection=true&name=testSimple1_shard2_replica_n4&action=CREATE&numShards=2&collection=testSimple1&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1629340 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr    x:testSimple1_shard1_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf&newCollection=true&name=testSimple1_shard1_replica_n2&action=CREATE&numShards=2&collection=testSimple1&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1629341 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr    x:testSimple1_shard1_replica_n2] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1629341 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr    x:testSimple1_shard2_replica_n6] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf&newCollection=true&name=testSimple1_shard2_replica_n6&action=CREATE&numShards=2&collection=testSimple1&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1629343 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr    x:testSimple1_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf&newCollection=true&name=testSimple1_shard1_replica_n1&action=CREATE&numShards=2&collection=testSimple1&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1630291 DEBUG (ScheduledTrigger-8462-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 3
   [junit4]   2> 1630346 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1630346 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1630360 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1630362 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 1630363 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.s.IndexSchema [testSimple1_shard2_replica_n4] Schema name=minimal
   [junit4]   2> 1630363 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.s.IndexSchema [testSimple1_shard2_replica_n6] Schema name=minimal
   [junit4]   2> 1630365 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 1630365 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard2_replica_n4' using configuration from collection testSimple1, trusted=true
   [junit4]   2> 1630366 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 1630366 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard2_replica_n6' using configuration from collection testSimple1, trusted=true
   [junit4]   2> 1630366 INFO  (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard2.replica_n4' (registry 'solr.core.testSimple1.shard2.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1630366 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard2.replica_n6' (registry 'solr.core.testSimple1.shard2.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1630367 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.s.IndexSchema [testSimple1_shard1_replica_n2] Schema name=minimal
   [junit4]   2> 1630378 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.s.IndexSchema [testSimple1_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 1630378 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 1630379 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard1_replica_n2' using configuration from collection testSimple1, trusted=true
   [junit4]   2> 1630379 INFO  (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard1.replica_n2' (registry 'solr.core.testSimple1.shard1.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1630379 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 1630379 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard1_replica_n1' using configuration from collection testSimple1, trusted=true
   [junit4]   2> 1630383 INFO  (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard1.replica_n1' (registry 'solr.core.testSimple1.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
   [junit4]   2> 1630385 INFO  (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory solr.hdf

[...truncated too long message...]

build/contrib/solr-dataimporthandler/test/temp/junit4-J0-20180623_003429_74715312066844850799510.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-dataimporthandler/test/temp/junit4-J1-20180623_003429_7479827180275547199107.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 567 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J1-20180623_003457_7593588562413959039049.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 6 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J0-20180623_003457_7592461396733423052156.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J2-20180623_003457_76214734783176774942234.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 576 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J2-20180623_003510_19510980844002221148259.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J0-20180623_003510_1958195709446057798390.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J1-20180623_003510_1952916833743249538977.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 674 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J2-20180623_003520_44115365425148263749297.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J0-20180623_003520_44116354916187501290760.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J1-20180623_003520_44113703101826179663573.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 570 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-prometheus-exporter/test/temp/junit4-J0-20180623_003550_60611854339242176193155.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-prometheus-exporter/test/temp/junit4-J1-20180623_003550_6069919291497190540503.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 606 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-uima/test/temp/junit4-J1-20180623_003611_4283766516359282848020.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-uima/test/temp/junit4-J0-20180623_003611_42815424760474818802855.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 557 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-velocity/test/temp/junit4-J0-20180623_003625_2359907790589491311819.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 34353 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22300/consoleText

[repro] Revision: 3b9d3a760a432b97aad2c08b2f778fa2344eb14a

[repro] Ant options: "-Dargs=-XX:-UseCompressedOops -XX:+UseConcMarkSweepGC"
[repro] Repro line:  ant test  -Dtestcase=HdfsAutoAddReplicasIntegrationTest -Dtests.method=testSimple -Dtests.seed=A483062C73879C9B -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ga -Dtests.timezone=Europe/Amsterdam -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1

[repro] ant clean

[...truncated 6 lines...]
[repro] Test suites by module:
[repro]    solr/core
[repro]       HdfsAutoAddReplicasIntegrationTest
[repro] ant compile-test

[...truncated 3305 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.HdfsAutoAddReplicasIntegrationTest" -Dtests.showOutput=onerror "-Dargs=-XX:-UseCompressedOops -XX:+UseConcMarkSweepGC" -Dtests.seed=A483062C73879C9B -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ga -Dtests.timezone=Europe/Amsterdam -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1

[...truncated 45 lines...]
   [junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J0-20180623_003921_9778145130587091982603.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 4 lines...]
   [junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J2-20180623_003921_97710617184633399344356.syserr
   [junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J2: EOF ----

[...truncated 4 lines...]
   [junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J1-20180623_003921_97713048276681065257358.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 23 lines...]
[repro] Failures:
[repro]   0/5 failed: org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest
[repro] Exiting with code 0

[...truncated 40 lines...]

[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-10.0.1) - Build # 22301 - Still Unstable!

Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22301/
Java: 64bit/jdk-10.0.1 -XX:-UseCompressedOops -XX:+UseSerialGC

1 tests failed.
FAILED:  org.apache.solr.cloud.cdcr.CdcrBidirectionalTest.testBiDir

Error Message:
Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]

Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
Caused by: java.lang.AssertionError
	at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
	at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
	at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
	at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
	at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 13442 lines...]
   [junit4] Suite: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest
   [junit4]   2> 726810 INFO  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/init-core-data-001
   [junit4]   2> 726813 WARN  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=25 numCloses=25
   [junit4]   2> 726813 INFO  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 726814 INFO  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0)
   [junit4]   2> 726821 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testBiDir
   [junit4]   2> 726821 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 1 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001
   [junit4]   2> 726822 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 726834 INFO  (Thread-2493) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 726834 INFO  (Thread-2493) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 726867 ERROR (Thread-2493) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 726934 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer start zk server on port:38873
   [junit4]   2> 726967 INFO  (zkConnectionManagerCallback-2451-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 726987 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
   [junit4]   2> 727006 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 727006 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 727006 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 727007 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@7c514370{/solr,null,AVAILABLE}
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@55d7ae79{SSL,[ssl, http/1.1]}{127.0.0.1:৩৯৩৪৯}
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.e.j.s.Server Started @৭২৭০৩৩ms
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=39349}
   [junit4]   2> 727008 ERROR (jetty-launcher-2448-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solrâ„¢ version 8.0.0
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 727008 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T04:08:53.265596Z
   [junit4]   2> 727034 INFO  (zkConnectionManagerCallback-2453-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727035 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 727144 INFO  (jetty-launcher-2448-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:38873/solr
   [junit4]   2> 727145 INFO  (zkConnectionManagerCallback-2457-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727146 INFO  (zkConnectionManagerCallback-2459-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727181 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 727182 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:39349_solr
   [junit4]   2> 727182 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.Overseer Overseer (id=72248175661154307-127.0.0.1:39349_solr-n_0000000000) starting
   [junit4]   2> 727186 INFO  (zkConnectionManagerCallback-2466-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727188 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
   [junit4]   2> 727188 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:39349_solr
   [junit4]   2> 727188 INFO  (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727198 INFO  (zkCallback-2465-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727210 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 727225 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727232 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727232 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727233 INFO  (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/.
   [junit4]   2> 727254 INFO  (zkConnectionManagerCallback-2469-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727258 INFO  (zkConnectionManagerCallback-2472-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727259 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 1 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001
   [junit4]   2> 727260 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 727260 INFO  (Thread-2503) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 727260 INFO  (Thread-2503) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 727267 ERROR (Thread-2503) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 727360 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer start zk server on port:44483
   [junit4]   2> 727361 INFO  (zkConnectionManagerCallback-2476-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727371 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
   [junit4]   2> 727373 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 727373 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 727373 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 727374 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@6ad5a77d{/solr,null,AVAILABLE}
   [junit4]   2> 727374 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@3cfe7871{SSL,[ssl, http/1.1]}{127.0.0.1:৪৩৭৯১}
   [junit4]   2> 727374 INFO  (jetty-launcher-2473-thread-1) [    ] o.e.j.s.Server Started @৭২৭৪০০ms
   [junit4]   2> 727374 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=43791}
   [junit4]   2> 727378 ERROR (jetty-launcher-2473-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solrâ„¢ version 8.0.0
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 727378 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-06-23T04:08:53.635561Z
   [junit4]   2> 727380 INFO  (zkConnectionManagerCallback-2478-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727381 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 727532 INFO  (jetty-launcher-2473-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44483/solr
   [junit4]   2> 727543 INFO  (zkConnectionManagerCallback-2482-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727546 INFO  (zkConnectionManagerCallback-2484-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727582 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 727582 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:43791_solr
   [junit4]   2> 727583 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.Overseer Overseer (id=72248175688417283-127.0.0.1:43791_solr-n_0000000000) starting
   [junit4]   2> 727590 INFO  (zkConnectionManagerCallback-2491-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727592 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44483/solr ready
   [junit4]   2> 727601 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:43791_solr
   [junit4]   2> 727618 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727623 INFO  (zkCallback-2490-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727632 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 727641 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727646 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727646 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 727647 INFO  (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/.
   [junit4]   2> 727674 INFO  (zkConnectionManagerCallback-2494-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727698 INFO  (zkConnectionManagerCallback-2497-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727703 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.CdcrBidirectionalTest cluster2 zkHost = 127.0.0.1:38873/solr
   [junit4]   2> 727703 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.CdcrBidirectionalTest cluster1 zkHost = 127.0.0.1:44483/solr
   [junit4]   2> 727710 INFO  (zkConnectionManagerCallback-2499-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727722 INFO  (zkConnectionManagerCallback-2503-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 727723 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 727724 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44483/solr ready
   [junit4]   2> 727738 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=cdcr-cluster1&maxShardsPerNode=2&name=cdcr-cluster1&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 727742 INFO  (OverseerThreadFactory-3404-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection cdcr-cluster1
   [junit4]   2> 727847 INFO  (OverseerStateUpdate-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster1",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"cdcr-cluster1_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:43791/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 727850 INFO  (OverseerStateUpdate-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster1",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"cdcr-cluster1_shard2_replica_n3",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:43791/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 728103 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr    x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster1_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 728104 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr    x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 728103 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr    x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster1_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 729144 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 729149 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.IndexSchema [cdcr-cluster1_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 729150 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 729150 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster1_shard1_replica_n1' using configuration from collection cdcr-cluster1, trusted=true
   [junit4]   2> 729151 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 729152 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.core.cdcr-cluster1.shard1.replica_n1' (registry 'solr.core.cdcr-cluster1.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 729154 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 729156 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [[cdcr-cluster1_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/cdcr-cluster1_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/./cdcr-cluster1_shard1_replica_n1/data/]
   [junit4]   2> 729156 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.IndexSchema [cdcr-cluster1_shard2_replica_n3] Schema name=minimal
   [junit4]   2> 729157 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 729157 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster1_shard2_replica_n3' using configuration from collection cdcr-cluster1, trusted=true
   [junit4]   2> 729158 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.core.cdcr-cluster1.shard2.replica_n3' (registry 'solr.core.cdcr-cluster1.shard2.replica_n3') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 729158 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 729158 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [[cdcr-cluster1_shard2_replica_n3] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/cdcr-cluster1_shard2_replica_n3], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/./cdcr-cluster1_shard2_replica_n3/data/]
   [junit4]   2> 729331 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 729331 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 729331 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 729331 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 729332 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 729332 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 729332 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 729332 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 729333 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@4cade66a[cdcr-cluster1_shard1_replica_n1] main]
   [junit4]   2> 729333 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@534a2f15[cdcr-cluster1_shard2_replica_n3] main]
   [junit4]   2> 729334 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster1
   [junit4]   2> 729334 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster1
   [junit4]   2> 729334 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster1
   [junit4]   2> 729334 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster1
   [junit4]   2> 729334 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster1/managed-schema
   [junit4]   2> 729334 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster1/managed-schema
   [junit4]   2> 729334 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 729334 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 729335 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 729335 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 729338 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.CdcrBufferStateManager Created znode /collections/cdcr-cluster1/cdcr/state/buffer
   [junit4]   2> 729340 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrProcessStateManager Created znode /collections/cdcr-cluster1/cdcr/state/process
   [junit4]   2> 729353 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034951234191360
   [junit4]   2> 729353 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034951234191360
   [junit4]   2> 729354 INFO  (searcherExecutor-3410-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard2_replica_n3 c:cdcr-cluster1 s:shard2 r:core_node4) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster1_shard2_replica_n3] Registered new searcher Searcher@534a2f15[cdcr-cluster1_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 729354 INFO  (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@4cade66a[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 729357 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard1 to Terms{values={core_node2=0}, version=0}
   [junit4]   2> 729358 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard2 to Terms{values={core_node4=0}, version=0}
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/ has no replicas
   [junit4]   2> 729359 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/ has no replicas
   [junit4]   2> 729360 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 729361 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster1:shard1
   [junit4]   2> 729364 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster1:shard2
   [junit4]   2> 729364 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/ shard1
   [junit4]   2> 729366 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/ shard2
   [junit4]   2> 729476 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 729477 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster1_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1374
   [junit4]   2> 729516 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 729517 INFO  (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster1_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1414
   [junit4]   2> 729519 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 729618 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/state.json] for collection [cdcr-cluster1] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 729744 INFO  (OverseerCollectionConfigSetProcessor-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 730519 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=cdcr-cluster1&maxShardsPerNode=2&name=cdcr-cluster1&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2} status=0 QTime=2781
   [junit4]   2> 730538 INFO  (zkConnectionManagerCallback-2507-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 730545 INFO  (zkConnectionManagerCallback-2511-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 730546 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 730560 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
   [junit4]   2> 730623 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=cdcr-cluster2&maxShardsPerNode=2&name=cdcr-cluster2&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 730639 INFO  (OverseerThreadFactory-3391-thread-1) [    ] o.a.s.c.a.c.CreateCollectionCmd Create collection cdcr-cluster2
   [junit4]   2> 730749 INFO  (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster2",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"cdcr-cluster2_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:39349/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 730751 INFO  (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"cdcr-cluster2",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"cdcr-cluster2_shard2_replica_n3",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"https://127.0.0.1:39349/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 730989 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr    x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster2_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 730990 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr    x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 731031 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr    x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster2_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
   [junit4]   2> 732008 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 732013 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.IndexSchema [cdcr-cluster2_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 732043 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 732044 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster2_shard1_replica_n1' using configuration from collection cdcr-cluster2, trusted=true
   [junit4]   2> 732044 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.core.cdcr-cluster2.shard1.replica_n1' (registry 'solr.core.cdcr-cluster2.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 732044 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 732044 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [[cdcr-cluster2_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/cdcr-cluster2_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/./cdcr-cluster2_shard1_replica_n1/data/]
   [junit4]   2> 732055 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
   [junit4]   2> 732063 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.IndexSchema [cdcr-cluster2_shard2_replica_n3] Schema name=minimal
   [junit4]   2> 732064 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 732064 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster2_shard2_replica_n3' using configuration from collection cdcr-cluster2, trusted=true
   [junit4]   2> 732065 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.core.cdcr-cluster2.shard2.replica_n3' (registry 'solr.core.cdcr-cluster2.shard2.replica_n3') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
   [junit4]   2> 732065 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 732065 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [[cdcr-cluster2_shard2_replica_n3] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/cdcr-cluster2_shard2_replica_n3], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/./cdcr-cluster2_shard2_replica_n3/data/]
   [junit4]   2> 732150 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 732151 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 732152 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 732152 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 732155 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@116811f2[cdcr-cluster2_shard1_replica_n1] main]
   [junit4]   2> 732156 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster2
   [junit4]   2> 732156 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster2
   [junit4]   2> 732156 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster2/managed-schema
   [junit4]   2> 732156 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 732157 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 732159 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrBufferStateManager Created znode /collections/cdcr-cluster2/cdcr/state/buffer
   [junit4]   2> 732165 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrProcessStateManager Created znode /collections/cdcr-cluster2/cdcr/state/process
   [junit4]   2> 732206 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034954225778688
   [junit4]   2> 732208 INFO  (searcherExecutor-3421-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster2_shard1_replica_n1] Registered new searcher Searcher@116811f2[cdcr-cluster2_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 732208 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
   [junit4]   2> 732208 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 732209 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 732209 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 732211 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster2/terms/shard1 to Terms{values={core_node2=0}, version=0}
   [junit4]   2> 732212 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@7910e119[cdcr-cluster2_shard2_replica_n3] main]
   [junit4]   2> 732212 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster2
   [junit4]   2> 732213 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster2
   [junit4]   2> 732213 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster2/managed-schema
   [junit4]   2> 732213 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 732214 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/ has no replicas
   [junit4]   2> 732215 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 732216 INFO  (zkCallback-2458-thread-2) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster2:shard1
   [junit4]   2> 732217 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/ shard1
   [junit4]   2> 732224 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034954244653056
   [junit4]   2> 732252 INFO  (searcherExecutor-3422-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster2_shard2_replica_n3] Registered new searcher Searcher@7910e119[cdcr-cluster2_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 732253 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster2/terms/shard2 to Terms{values={core_node4=0}, version=0}
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/ has no replicas
   [junit4]   2> 732255 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 732256 INFO  (zkCallback-2458-thread-2) [    ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster2:shard2
   [junit4]   2> 732258 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/ shard2
   [junit4]   2> 732360 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 732361 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster2_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1330
   [junit4]   2> 732368 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 732369 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster2_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1380
   [junit4]   2> 732373 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 732470 INFO  (zkCallback-2458-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster2/state.json] for collection [cdcr-cluster2] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 732662 INFO  (OverseerCollectionConfigSetProcessor-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 733373 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=cdcr-cluster2&maxShardsPerNode=2&name=cdcr-cluster2&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2} status=0 QTime=2750
   [junit4]   2> 733390 INFO  (zkConnectionManagerCallback-2518-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 733391 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 733391 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
   [junit4]   2> 733421 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 733421 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=3
   [junit4]   2> 733443 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 733443 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733444 INFO  (qtp185556678-10496) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=45
   [junit4]   2> 733445 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard1
   [junit4]   2> 733445 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2, shard: shard1
   [junit4]   2> 733447 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Submitting bootstrap task to executor
   [junit4]   2> 733451 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2 shard: shard1 leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/
   [junit4]   2> 733461 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrProcessStateManager The CDCR process state has changed: WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/cdcr/state/process @ cdcr-cluster1:shard2
   [junit4]   2> 733461 INFO  (zkCallback-2483-thread-2) [    ] o.a.s.h.CdcrProcessStateManager The CDCR process state has changed: WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/cdcr/state/process @ cdcr-cluster1:shard1
   [junit4]   2> 733462 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrProcessStateManager Received new CDCR process state from watcher: STARTED @ cdcr-cluster1:shard2
   [junit4]   2> 733462 INFO  (zkCallback-2483-thread-2) [    ] o.a.s.h.CdcrProcessStateManager Received new CDCR process state from watcher: STARTED @ cdcr-cluster1:shard1
   [junit4]   2> 733462 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/cdcr params={qt=/cdcr&_stateVer_=cdcr-cluster1:4&action=start&wt=javabin&version=2} status=0 QTime=85
   [junit4]   2> 733514 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={qt=/cdcr&masterUrl=https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/&action=BOOTSTRAP&wt=javabin&version=2} status=0 QTime=5
   [junit4]   2> 733516 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733518 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager CDCR bootstrap running for 1 seconds, sleeping for 2000 ms
   [junit4]   2> 733526 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 733534 INFO  (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034955618287616,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 733542 INFO  (zkConnectionManagerCallback-2523-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 733542 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 733543 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
   [junit4]   2> 733554 INFO  (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 733555 INFO  (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 733555 INFO  (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={waitSearcher=true&openSearcher=false&commit=true&softCommit=false&commit_end_point=true&wt=javabin&version=2} status=0 QTime=22
   [junit4]   2> 733585 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 733585 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733609 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 733609 INFO  (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733610 INFO  (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=64
   [junit4]   2> 733610 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard2
   [junit4]   2> 733610 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2, shard: shard2
   [junit4]   2> 733610 INFO  (zkCallback-2483-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Submitting bootstrap task to executor
   [junit4]   2> 733615 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2 shard: shard2 leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/
   [junit4]   2> 733622 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/replication params={qt=/replication&wt=javabin&version=2&command=indexversion} status=0 QTime=0
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Master's generation: 1
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Master's version: 0
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Slave's generation: 1
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Slave's version: 0
   [junit4]   2> 733622 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher New index in Master. Deleting mine...
   [junit4]   2> 733627 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@860e11[cdcr-cluster2_shard1_replica_n1] main]
   [junit4]   2> 733628 INFO  (searcherExecutor-3421-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster2_shard1_replica_n1] Registered new searcher Searcher@860e11[cdcr-cluster2_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 733628 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler No replay needed.
   [junit4]   2> 733628 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={qt=/cdcr&masterUrl=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&action=BOOTSTRAP&wt=javabin&version=2} status=0 QTime=9
   [junit4]   2> 733629 INFO  (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 733630 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager CDCR bootstrap running for 1 seconds, sleeping for 2000 ms
   [junit4]   2> 733630 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 733641 INFO  (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034955730485248,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 733646 INFO  (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 733646 INFO  (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 733646 INFO  (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={waitSearcher=true&openSearcher=false&commit=true&softCommit=false&commit_end_point=true&wt=javabin&version=2} status=0 QTime=6
   [junit4]   2> 733650 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/replication params={qt=/replication&wt=javabin&version=2&command=indexversion} status=0 QTime=0
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Master's generation: 1
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Master's version: 0
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Slave's generation: 1
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Slave's version: 0
   [junit4]   2> 733650 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher New index in Master. Deleting mine...
   [junit4]   2> 733651 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@3853640e[cdcr-cluster2_shard2_replica_n3] main]
   [junit4]   2> 733652 INFO  (searcherExecutor-3422-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster2_shard2_replica_n3] Registered new searcher Searcher@3853640e[cdcr-cluster2_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 733652 INFO  (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler No replay needed.
   [junit4]   2> 735463 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.CdcrBidirectionalTest Adding 10 docs with commit=true, numDocs=100
   [junit4]   2> 735469 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard1 to Terms{values={core_node2=1}, version=1}
   [junit4]   2> 735469 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=3
   [junit4]   2> 735469 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard2 to Terms{values={core_node4=1}, version=1}
   [junit4]   2> 735469 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=5
   [junit4]   2> 735474 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957652525056,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 735474 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@62bc2cb1 commitCommandVersion:1604034957652525056
   [junit4]   2> 735478 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957656719360,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 735478 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@6850d084 commitCommandVersion:1604034957656719360
   [junit4]   2> 735518 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735519 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager CDCR bootstrap successful in 3 seconds
   [junit4]   2> 735523 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 735523 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735525 INFO  (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 735525 INFO  (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735528 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=9
   [junit4]   2> 735529 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard1
   [junit4]   2> 735529 INFO  (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Bootstrap successful, giving the go-ahead to replicator
   [junit4]   2> 735630 INFO  (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735631 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager CDCR bootstrap successful in 3 seconds
   [junit4]   2> 735637 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 735637 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
   [junit4]   2> 735637 INFO  (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735637 INFO  (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1]  webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
   [junit4]   2> 735638 INFO  (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3]  webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=6
   [junit4]   2> 735638 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard2
   [junit4]   2> 735638 INFO  (cdcr-bootstrap-status-2519-thread-1) [    ] o.a.s.h.CdcrReplicatorManager Bootstrap successful, giving the go-ahead to replicator
   [junit4]   2> 735653 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@601f19c[cdcr-cluster1_shard1_replica_n1] main]
   [junit4]   2> 735654 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 735656 INFO  (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@601f19c[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C43)))}
   [junit4]   2> 735656 INFO  (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false} status=0 QTime=182
   [junit4]   2> 735667 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@6fb4ca9f[cdcr-cluster1_shard2_replica_n3] main]
   [junit4]   2> 735668 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 735669 INFO  (searcherExecutor-3410-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard2_replica_n3 c:cdcr-cluster1 s:shard2 r:core_node4) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster1_shard2_replica_n3] Registered new searcher Searcher@6fb4ca9f[cdcr-cluster1_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C57)))}
   [junit4]   2> 735669 INFO  (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false} status=0 QTime=191
   [junit4]   2> 735669 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&waitSearcher=true&commit=true&softCommit=false&wt=javabin&version=2} status=0 QTime=199
   [junit4]   2> 735670 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.c.CdcrBidirectionalTest Adding 10 docs with commit=true, numDocs=200
   [junit4]   2> 735680 INFO  (qtp804107552-10610) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=9
   [junit4]   2> 735683 INFO  (qtp804107552-10609) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3]  webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=6
   [junit4]   2> 735684 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957872726016,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 735684 INFO  (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@6850d084 commitCommandVersion:1604034957872726016
   [junit4]   2> 735686 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957874823168,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 735686 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@62bc2cb1 commitCommandVersion:1604034957874823168
   [junit4]   2> 735972 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@3f3de49e[cdcr-cluster1_shard1_replica_n1] main]
   [junit4]   2> 735973 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 735974 INFO  (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@3f3de49e[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C43) Uninverting(_1(8.0.0):C54)))}
   [junit4]   2> 735974 INFO  (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1]  webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}

[...truncated too long message...]

try=solr.collection.cdcr-cluster2.shard2.leader, tag=18455e69
   [junit4]   2> 760651 INFO  (coreCloseExecutor-3451-thread-2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Solr core is being closed - shutting down CDCR handler @ cdcr-cluster2:shard2
   [junit4]   2> 760731 INFO  (jetty-closer-2449-thread-1) [    ] o.a.s.c.Overseer Overseer (id=72248175661154307-127.0.0.1:39349_solr-n_0000000000) closing
   [junit4]   2> 760731 INFO  (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.Overseer Overseer Loop exiting : 127.0.0.1:39349_solr
   [junit4]   2> 760731 WARN  (OverseerAutoScalingTriggerThread-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread woken up but we are closed, exiting.
   [junit4]   2> 760732 INFO  (jetty-closer-2449-thread-1) [    ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@7c514370{/solr,null,UNAVAILABLE}
   [junit4]   2> 760732 INFO  (jetty-closer-2449-thread-1) [    ] o.e.j.s.session node0 Stopped scavenging
   [junit4]   2> 760733 ERROR (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 760733 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:38873 38873
   [junit4]   2> 760778 INFO  (Thread-2493) [    ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:38873 38873
   [junit4]   2> 760778 WARN  (Thread-2493) [    ] o.a.s.c.ZkTestServer Watch limit violations: 
   [junit4]   2> Maximum concurrent create/delete watches above limit:
   [junit4]   2> 
   [junit4]   2> 	5	/solr/aliases.json
   [junit4]   2> 	3	/solr/collections/cdcr-cluster2/terms/shard2
   [junit4]   2> 	3	/solr/collections/cdcr-cluster2/terms/shard1
   [junit4]   2> 	2	/solr/configs/cdcr-cluster2/managed-schema
   [junit4]   2> 
   [junit4]   2> Maximum concurrent data watches above limit:
   [junit4]   2> 
   [junit4]   2> 	6	/solr/collections/cdcr-cluster2/cdcr/state/process
   [junit4]   2> 	5	/solr/clusterstate.json
   [junit4]   2> 	5	/solr/clusterprops.json
   [junit4]   2> 	3	/solr/collections/cdcr-cluster2/state.json
   [junit4]   2> 	2	/solr/collections/cdcr-cluster2/cdcr/state/buffer
   [junit4]   2> 
   [junit4]   2> Maximum concurrent children watches above limit:
   [junit4]   2> 
   [junit4]   2> 	5	/solr/live_nodes
   [junit4]   2> 	5	/solr/collections
   [junit4]   2> 
   [junit4]   2> 760779 INFO  (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [    ] o.a.s.SolrTestCaseJ4 ###Ending testBiDir
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=CdcrBidirectionalTest -Dtests.method=testBiDir -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8
   [junit4] ERROR   34.0s J0 | CdcrBidirectionalTest.testBiDir <<<
   [junit4]    > Throwable #1: com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
   [junit4]    > Caused by: java.lang.AssertionError
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
   [junit4]    > 	at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
   [junit4]    > 	at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
   [junit4]    > 	at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
   [junit4]    > 	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
   [junit4]    > 	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
   [junit4]    > 	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
   [junit4]    > 	at java.base/java.lang.Thread.run(Thread.java:844)Throwable #2: com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10729, name=cdcr-replicator-3446-thread-2, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
   [junit4]    > Caused by: java.lang.AssertionError
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
   [junit4]    > 	at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
   [junit4]    > 	at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
   [junit4]    > 	at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
   [junit4]    > 	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
   [junit4]    > 	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
   [junit4]    > 	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
   [junit4]    > 	at java.base/java.lang.Thread.run(Thread.java:844)
   [junit4]   2> 760780 INFO  (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [    ] o.a.s.SolrTestCaseJ4 ------------------------------------------------------- Done waiting for tracked resources to be released
   [junit4]   2> Jun 23, 2018 4:09:27 AM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
   [junit4]   2> WARNING: Will linger awaiting termination of 5 leaked thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene70), sim=Asserting(org.apache.lucene.search.similarities.AssertingSimilarity@8d81519), locale=bn-IN, timezone=Asia/Tomsk
   [junit4]   2> NOTE: Linux 4.13.0-41-generic amd64/Oracle Corporation 10.0.1 (64-bit)/cpus=8,threads=1,free=149515200,total=518979584
   [junit4]   2> NOTE: All tests run in this JVM: [TestNestedDocsSort, TestMaxScoreQueryParser, URLClassifyProcessorTest, ZkCLITest, TestHttpShardHandlerFactory, SimpleCollectionCreateDeleteTest, TestClusterStateMutator, TestMinMaxOnMultiValuedField, ZkNodePropsTest, TestSort, TestComplexPhraseLeadingWildcard, TestHdfsBackupRestoreCore, TestSystemIdResolver, CursorPagingTest, TestSolrCloudWithDelegationTokens, ChaosMonkeyNothingIsSafeTest, TestBulkSchemaConcurrent, TestCollationFieldDocValues, DeleteLastCustomShardedReplicaTest, TestNumericRangeQuery32, TestRTGBase, TestManagedSchemaAPI, FileUtilsTest, BlockJoinFacetSimpleTest, TestShortCircuitedRequests, TestJettySolrRunner, BigEndianAscendingWordSerializerTest, TestCSVLoader, TestExecutePlanAction, StatsComponentTest, AtomicUpdateProcessorFactoryTest, RegexBytesRefFilterTest, HdfsUnloadDistributedZkTest, HealthCheckHandlerTest, SolrCoreMetricManagerTest, TestRequestStatusCollectionAPI, BadCopyFieldTest, SolrXmlInZkTest, TestRemoteStreaming, IndexBasedSpellCheckerTest, TestStressCloudBlindAtomicUpdates, TestPrepRecovery, TestUtilizeNode, RequestHandlersTest, TestObjectReleaseTracker, TestDistribIDF, HdfsAutoAddReplicasIntegrationTest, SchemaApiFailureTest, FieldAnalysisRequestHandlerTest, SystemInfoHandlerTest, DistributedQueryElevationComponentTest, BJQParserTest, MBeansHandlerTest, CollectionsAPIAsyncDistributedZkTest, DocumentAnalysisRequestHandlerTest, AutoScalingHandlerTest, PKIAuthenticationIntegrationTest, TestRawTransformer, XmlInterpolationTest, TestSolrDeletionPolicy1, TestExportWriter, TestManagedStopFilterFactory, TestQuerySenderNoQuery, MetricsHistoryHandlerTest, TestDocBasedVersionConstraints, NodeMarkersRegistrationTest, FullSolrCloudDistribCmdsTest, TestCloudDeleteByQuery, TestSolrCLIRunExample, ShufflingReplicaListTransformerTest, AssignTest, BlobRepositoryCloudTest, TestCrossCoreJoin, HDFSCollectionsAPITest, HttpSolrCallGetCoreTest, TestLeaderElectionWithEmptyReplica, TestLuceneMatchVersion, TestLegacyNumericUtils, HdfsChaosMonkeyNothingIsSafeTest, ConfigureRecoveryStrategyTest, TestIBSimilarityFactory, TestEmbeddedSolrServerAdminHandler, HdfsSyncSliceTest, CoreMergeIndexesAdminHandlerTest, CustomCollectionTest, TestConfig, SimpleFacetsTest, OverseerStatusTest, HighlighterConfigTest, RollingRestartTest, TestSolrQueryParser, TestDocTermOrdsUninvertLimit, TestDistributedMap, SolrCloudReportersTest, TestXmlQParser, TestIndexSearcher, TestGroupingSearch, TestCustomStream, SharedFSAutoReplicaFailoverTest, RAMDirectoryFactoryTest, TestJsonRequest, ProtectedTermFilterFactoryTest, TestJsonFacetRefinement, CdcrBidirectionalTest]
   [junit4] Completed [310/821 (1!)] on J0 in 34.51s, 1 test, 1 error <<< FAILURES!

[...truncated 45707 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22301/consoleText

[repro] Revision: 3b9d3a760a432b97aad2c08b2f778fa2344eb14a

[repro] Ant options: "-Dargs=-XX:-UseCompressedOops -XX:+UseSerialGC"
[repro] Repro line:  ant test  -Dtestcase=CdcrBidirectionalTest -Dtests.method=testBiDir -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8

[repro] ant clean

[...truncated 6 lines...]
[repro] Test suites by module:
[repro]    solr/core
[repro]       CdcrBidirectionalTest
[repro] ant compile-test

[...truncated 3305 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.CdcrBidirectionalTest" -Dtests.showOutput=onerror "-Dargs=-XX:-UseCompressedOops -XX:+UseSerialGC" -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8

[...truncated 76 lines...]
[repro] Failures:
[repro]   0/5 failed: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest
[repro] Exiting with code 0

[...truncated 40 lines...]