You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2018/06/23 00:45:26 UTC
[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-10) - Build # 22300 -
Still Unstable!
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22300/
Java: 64bit/jdk-10 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC
1 tests failed.
FAILED: org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest.testSimple
Error Message:
Waiting for collection testSimple2 null Live Nodes: [127.0.0.1:34223_solr, 127.0.0.1:45039_solr] Last available state: DocCollection(testSimple2//collections/testSimple2/state.json/21)={ "pullReplicas":"0", "replicationFactor":"2", "shards":{ "shard1":{ "range":"80000000-ffffffff", "state":"active", "replicas":{ "core_node3":{ "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/", "base_url":"http://127.0.0.1:34223/solr", "node_name":"127.0.0.1:34223_solr", "type":"NRT", "force_set_state":"false", "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/tlog", "core":"testSimple2_shard1_replica_n1", "shared_storage":"true", "state":"active", "leader":"true"}, "core_node5":{ "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/", "base_url":"http://127.0.0.1:39879/solr", "node_name":"127.0.0.1:39879_solr", "type":"NRT", "force_set_state":"false", "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/tlog", "core":"testSimple2_shard1_replica_n2", "shared_storage":"true", "state":"down"}}}, "shard2":{ "range":"0-7fffffff", "state":"active", "replicas":{ "core_node6":{ "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/", "base_url":"http://127.0.0.1:34223/solr", "node_name":"127.0.0.1:34223_solr", "type":"NRT", "force_set_state":"false", "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/tlog", "core":"testSimple2_shard2_replica_n4", "shared_storage":"true", "state":"active", "leader":"true"}, "core_node8":{ "dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/", "base_url":"http://127.0.0.1:39879/solr", "node_name":"127.0.0.1:39879_solr", "type":"NRT", "force_set_state":"false", "ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/tlog", "core":"testSimple2_shard2_replica_n7", "shared_storage":"true", "state":"down"}}}}, "router":{"name":"compositeId"}, "maxShardsPerNode":"2", "autoAddReplicas":"true", "nrtReplicas":"2", "tlogReplicas":"0"}
Stack Trace:
java.lang.AssertionError: Waiting for collection testSimple2
null
Live Nodes: [127.0.0.1:34223_solr, 127.0.0.1:45039_solr]
Last available state: DocCollection(testSimple2//collections/testSimple2/state.json/21)={
"pullReplicas":"0",
"replicationFactor":"2",
"shards":{
"shard1":{
"range":"80000000-ffffffff",
"state":"active",
"replicas":{
"core_node3":{
"dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/",
"base_url":"http://127.0.0.1:34223/solr",
"node_name":"127.0.0.1:34223_solr",
"type":"NRT",
"force_set_state":"false",
"ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node3/data/tlog",
"core":"testSimple2_shard1_replica_n1",
"shared_storage":"true",
"state":"active",
"leader":"true"},
"core_node5":{
"dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/",
"base_url":"http://127.0.0.1:39879/solr",
"node_name":"127.0.0.1:39879_solr",
"type":"NRT",
"force_set_state":"false",
"ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node5/data/tlog",
"core":"testSimple2_shard1_replica_n2",
"shared_storage":"true",
"state":"down"}}},
"shard2":{
"range":"0-7fffffff",
"state":"active",
"replicas":{
"core_node6":{
"dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/",
"base_url":"http://127.0.0.1:34223/solr",
"node_name":"127.0.0.1:34223_solr",
"type":"NRT",
"force_set_state":"false",
"ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node6/data/tlog",
"core":"testSimple2_shard2_replica_n4",
"shared_storage":"true",
"state":"active",
"leader":"true"},
"core_node8":{
"dataDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/",
"base_url":"http://127.0.0.1:39879/solr",
"node_name":"127.0.0.1:39879_solr",
"type":"NRT",
"force_set_state":"false",
"ulogDir":"hdfs://localhost.localdomain:46029/data/testSimple2/core_node8/data/tlog",
"core":"testSimple2_shard2_replica_n7",
"shared_storage":"true",
"state":"down"}}}},
"router":{"name":"compositeId"},
"maxShardsPerNode":"2",
"autoAddReplicas":"true",
"nrtReplicas":"2",
"tlogReplicas":"0"}
at __randomizedtesting.SeedInfo.seed([A483062C73879C9B:9C3022D25474484A]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.apache.solr.cloud.SolrCloudTestCase.waitForState(SolrCloudTestCase.java:278)
at org.apache.solr.cloud.autoscaling.AutoAddReplicasIntegrationTest.testSimple(AutoAddReplicasIntegrationTest.java:129)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:564)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at java.base/java.lang.Thread.run(Thread.java:844)
Build Log:
[...truncated 1871 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J0-20180622_233120_8438624310112548928347.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J2-20180622_233120_843109218256741183964.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 5 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/core/test/temp/junit4-J1-20180622_233120_84215878213975654336568.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 298 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J1-20180622_234003_82713161516520257880618.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J2-20180622_234003_82211985366498351429167.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/test-framework/test/temp/junit4-J0-20180622_234003_8187629202302504117049.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 1080 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J0-20180622_234142_9048161528471907446887.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J2-20180622_234142_90416258816079129641780.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/common/test/temp/junit4-J1-20180622_234142_90412005419182290644372.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 258 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J0-20180622_234332_4914664188095482509082.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J2-20180622_234332_4922443722992351897655.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/icu/test/temp/junit4-J1-20180622_234332_4914364283643442390734.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 250 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J0-20180622_234345_5569241898986794068605.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J2-20180622_234345_55612839679531665619473.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 6 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/kuromoji/test/temp/junit4-J1-20180622_234345_55614100303459436628584.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 162 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J1-20180622_234408_1182183404860930597018.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J0-20180622_234408_11817221358498753266703.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/morfologik/test/temp/junit4-J2-20180622_234408_1191547639190121044076.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 210 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J0-20180622_234412_40714371405619675624954.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J2-20180622_234412_4078848850315630926441.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/nori/test/temp/junit4-J1-20180622_234412_4079975389867110683768.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 174 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J1-20180622_234421_51010637845793649127342.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J0-20180622_234421_5103569472874704098846.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/opennlp/test/temp/junit4-J2-20180622_234421_51016436102187679443010.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 176 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J2-20180622_234424_31610752690720732553073.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J0-20180622_234424_3163920468419256368342.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/phonetic/test/temp/junit4-J1-20180622_234424_31613507803586776163227.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 161 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/smartcn/test/temp/junit4-J0-20180622_234434_0177751930883938230811.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/smartcn/test/temp/junit4-J1-20180622_234434_01715792206888438466742.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 162 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J1-20180622_234443_3098926868024307710344.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J2-20180622_234443_30910390262105719088600.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/stempel/test/temp/junit4-J0-20180622_234443_3078828821911177442895.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 177 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/analysis/uima/test/temp/junit4-J0-20180622_234446_5954096128335448572284.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 153 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J1-20180622_234455_65114409003380652460182.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 5 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J2-20180622_234455_6514623326656775997862.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 19 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/backward-codecs/test/temp/junit4-J0-20180622_234455_6517020648835771442894.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 1391 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J2-20180622_234519_1932894334289275466381.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J1-20180622_234519_1932449911797630396956.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/benchmark/test/temp/junit4-J0-20180622_234519_1936352665634460400351.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 253 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J2-20180622_234527_4918016004016563548263.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J0-20180622_234527_49118385300419579868577.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/classification/test/temp/junit4-J1-20180622_234527_4915089559484994378250.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 262 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J1-20180622_234539_34511673298149119415382.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J0-20180622_234539_3442056690456979073956.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/codecs/test/temp/junit4-J2-20180622_234539_3456681254644361673058.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 231 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J1-20180622_234754_3966129908289707100345.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J2-20180622_234754_39610786966458196672386.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/demo/test/temp/junit4-J0-20180622_234754_39611330166380200542034.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 176 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J1-20180622_234757_3282744005959643641695.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J2-20180622_234757_328184188593405154770.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/expressions/test/temp/junit4-J0-20180622_234757_32813663791879804413349.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 235 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J1-20180622_234803_6451920093429510786067.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J2-20180622_234803_6456139679212212410020.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/facet/test/temp/junit4-J0-20180622_234803_64513176520531889858516.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 185 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J2-20180622_234842_66210931058644992560173.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J1-20180622_234842_66216254532774656442410.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/grouping/test/temp/junit4-J0-20180622_234842_6625491280472342315484.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 255 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J1-20180622_234900_48815603692318856114647.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J2-20180622_234900_48816107964350347376247.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 6 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/highlighter/test/temp/junit4-J0-20180622_234900_48816853337107767940056.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 163 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J0-20180622_234916_9509003729064799529499.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J2-20180622_234916_9509833749328796722833.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 6 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/join/test/temp/junit4-J1-20180622_234916_9501938938754275132751.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 155 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/memory/test/temp/junit4-J0-20180622_234945_85111956510346757286560.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/memory/test/temp/junit4-J1-20180622_234945_8526925569142241457753.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 189 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J2-20180622_234957_89716020023611267993152.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J1-20180622_234957_89713374075461271625175.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 6 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/misc/test/temp/junit4-J0-20180622_234957_897995069218042617454.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 304 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J1-20180622_235015_07816445064399041937754.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 6 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J2-20180622_235015_07817300623641250331463.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queries/test/temp/junit4-J0-20180622_235015_0777051079486333859241.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 229 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J2-20180622_235023_9761355228138200064238.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J0-20180622_235023_9765821768678419074163.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/queryparser/test/temp/junit4-J1-20180622_235023_9764016929074376670969.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 212 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J1-20180622_235031_70213674486619839082375.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J0-20180622_235031_70211591443367436951168.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/replicator/test/temp/junit4-J2-20180622_235031_70217646439452202035793.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 227 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J0-20180622_235044_36518075386306318393681.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J2-20180622_235044_3655371863721534862338.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 7 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/sandbox/test/temp/junit4-J1-20180622_235044_3653195148306886251101.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 298 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J0-20180622_235133_51311242690447766539259.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 6 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J2-20180622_235133_5131023607900679846491.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial-extras/test/temp/junit4-J1-20180622_235133_5137580074882706317880.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 153 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial/test/temp/junit4-J0-20180622_235158_4753421684692232253190.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 181 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J0-20180622_235200_23017643266495511437138.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 6 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J1-20180622_235200_2306779390704051548979.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 5 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/spatial3d/test/temp/junit4-J2-20180622_235200_23015255487659489383656.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 264 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J2-20180622_235320_44513781554556465979598.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J0-20180622_235320_4444365517420477492284.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/lucene/build/suggest/test/temp/junit4-J1-20180622_235320_4449138489014132900971.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3929 lines...]
[junit4] Suite: org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest
[junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/init-core-data-001
[junit4] 2> 1628046 WARN (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=1 numCloses=1
[junit4] 2> 1628046 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.SolrTestCaseJ4 Using TrieFields (NUMERIC_POINTS_SYSPROP=false) w/NUMERIC_DOCVALUES_SYSPROP=true
[junit4] 2> 1628047 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0)
[junit4] 2> 1628047 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> 1628047 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 3 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001
[junit4] 2> 1628047 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 1628048 INFO (Thread-4546) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 1628048 INFO (Thread-4546) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 1628049 ERROR (Thread-4546) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 1628148 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.c.ZkTestServer start zk server on port:36131
[junit4] 2> 1628149 INFO (zkConnectionManagerCallback-10100-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628151 INFO (jetty-launcher-10097-thread-2) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
[junit4] 2> 1628151 INFO (jetty-launcher-10097-thread-3) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
[junit4] 2> 1628151 INFO (jetty-launcher-10097-thread-1) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10+46
[junit4] 2> 1628151 INFO (jetty-launcher-10097-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1628151 INFO (jetty-launcher-10097-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1628151 INFO (jetty-launcher-10097-thread-1) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@4ebf9fd8{/solr,null,AVAILABLE}
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@636e248d{/solr,null,AVAILABLE}
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@3a13410e{/solr,null,AVAILABLE}
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.e.j.s.AbstractConnector Started ServerConnector@295b2198{HTTP/1.1,[http/1.1]}{127.0.0.1:45039}
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@dd199bf{HTTP/1.1,[http/1.1]}{127.0.0.1:39879}
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.e.j.s.Server Started @1628177ms
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.e.j.s.AbstractConnector Started ServerConnector@7ef44222{HTTP/1.1,[http/1.1]}{127.0.0.1:34223}
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-1) [ ] o.e.j.s.Server Started @1628177ms
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.e.j.s.Server Started @1628177ms
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=45039}
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=39879}
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=34223}
[junit4] 2> 1628152 ERROR (jetty-launcher-10097-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1628152 ERROR (jetty-launcher-10097-thread-3) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1628152 ERROR (jetty-launcher-10097-thread-2) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 8.0.0
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 8.0.0
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solr? version 8.0.0
[junit4] 2> 1628153 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1628152 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1628153 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1628153 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 1628153 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-23T00:21:31.684042Z
[junit4] 2> 1628153 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1628153 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 1628153 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-23T00:21:31.684074Z
[junit4] 2> 1628153 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-23T00:21:31.684084Z
[junit4] 2> 1628153 INFO (zkConnectionManagerCallback-10104-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628153 INFO (zkConnectionManagerCallback-10106-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628153 INFO (zkConnectionManagerCallback-10103-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628154 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1628154 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1628154 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 1628155 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
[junit4] 2> 1628155 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 1628155 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
[junit4] 2> 1628155 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 1628155 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
[junit4] 2> 1628155 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
[junit4] 2> 1628156 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
[junit4] 2> 1628156 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
[junit4] 2> 1628157 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156, but no JMX reporters were configured - adding default JMX reporter.
[junit4] 2> 1628243 INFO (jetty-launcher-10097-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
[junit4] 2> 1628244 INFO (zkConnectionManagerCallback-10110-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628245 INFO (zkConnectionManagerCallback-10112-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628251 INFO (jetty-launcher-10097-thread-3) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
[junit4] 2> 1628260 INFO (zkConnectionManagerCallback-10118-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628265 INFO (zkConnectionManagerCallback-10120-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628269 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1628269 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:45039_solr
[junit4] 2> 1628270 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.c.Overseer Overseer (id=72247281650171911-127.0.0.1:45039_solr-n_0000000000) starting
[junit4] 2> 1628281 INFO (zkConnectionManagerCallback-10127-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628282 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
[junit4] 2> 1628284 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:45039_solr
[junit4] 2> 1628285 INFO (zkCallback-10126-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1628285 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread Adding .auto_add_replicas and .scheduled_maintenance triggers
[junit4] 2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 1
[junit4] 2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 1, lastZnodeVersion -1
[junit4] 2> 1628286 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 1
[junit4] 2> 1628289 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:45039_solr]
[junit4] 2> 1628289 DEBUG (ScheduledTrigger-8462-thread-1) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
[junit4] 2> 1628292 INFO (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1628296 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
[junit4] 2> 1628296 DEBUG (OverseerAutoScalingTriggerThread-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 1, lastZnodeVersion 1
[junit4] 2> 1628297 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 1628297 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1628298 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
[junit4] 2> 1628298 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:39879_solr
[junit4] 2> 1628299 INFO (zkCallback-10119-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1628299 INFO (zkCallback-10126-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1628305 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1628313 INFO (zkCallback-10111-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
[junit4] 2> 1628317 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628317 INFO (zkConnectionManagerCallback-10133-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628318 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1628318 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
[junit4] 2> 1628318 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1628323 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628323 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628324 INFO (jetty-launcher-10097-thread-3) [n:127.0.0.1:45039_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node3/.
[junit4] 2> 1628330 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628336 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628336 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628337 INFO (jetty-launcher-10097-thread-1) [n:127.0.0.1:39879_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node1/.
[junit4] 2> 1628408 INFO (jetty-launcher-10097-thread-2) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:36131/solr
[junit4] 2> 1628409 INFO (zkConnectionManagerCallback-10138-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628410 INFO (zkConnectionManagerCallback-10140-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628413 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
[junit4] 2> 1628414 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 1628415 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
[junit4] 2> 1628415 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34223_solr
[junit4] 2> 1628415 INFO (zkCallback-10119-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1628415 INFO (zkCallback-10126-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1628415 INFO (zkCallback-10111-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1628415 INFO (zkCallback-10139-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1628415 INFO (zkCallback-10132-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
[junit4] 2> 1628424 INFO (zkConnectionManagerCallback-10147-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628425 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 1628425 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
[junit4] 2> 1628426 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 1628435 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628440 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628440 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1628441 INFO (jetty-launcher-10097-thread-2) [n:127.0.0.1:34223_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-001/node2/.
[junit4] 2> 1628457 INFO (zkConnectionManagerCallback-10150-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628459 INFO (zkConnectionManagerCallback-10155-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 1628459 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
[junit4] 2> 1628460 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:36131/solr ready
[junit4] 1> Formatting using clusterid: testClusterID
[junit4] 2> 1628518 WARN (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
[junit4] 2> 1628524 WARN (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1628525 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1628531 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_localdomain_33425_hdfs____.4bsgsv/webapp
[junit4] 2> 1628601 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost.localdomain:33425
[junit4] 2> 1628678 WARN (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1628679 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1628683 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_34005_datanode____3r5l3v/webapp
[junit4] 2> 1628761 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:34005
[junit4] 2> 1628785 WARN (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
[junit4] 2> 1628786 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log jetty-6.1.26
[junit4] 2> 1628789 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log Extract jar:file:/home/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_32985_datanode____.yy7xwi/webapp
[junit4] 2> 1628864 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data2/]] heartbeating to localhost.localdomain/127.0.0.1:46029) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 1628869 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xa4871f7444f09: from storage DS-53fc62e1-2ec1-4aa4-a0a1-da327bd2c668 node DatanodeRegistration(127.0.0.1:40969, datanodeUuid=6fa86369-873f-4f5b-b41f-1e017be7aac9, infoPort=35381, infoSecurePort=0, ipcPort=35747, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 1628869 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xa4871f7444f09: from storage DS-ccdddced-bb98-4bc9-9df8-d7f561692c23 node DatanodeRegistration(127.0.0.1:40969, datanodeUuid=6fa86369-873f-4f5b-b41f-1e017be7aac9, infoPort=35381, infoSecurePort=0, ipcPort=35747, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 1628872 INFO (SUITE-HdfsAutoAddReplicasIntegrationTest-seed#[A483062C73879C9B]-worker) [ ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985
[junit4] 2> 1628873 WARN (1512157112@qtp-148498253-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985) [ ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is false. Rechecking.
[junit4] 2> 1628883 WARN (1512157112@qtp-148498253-1 - Acceptor0 HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:32985) [ ] o.a.h.h.HttpServer2 HttpServer Acceptor: isRunning is true
[junit4] 2> 1628968 ERROR (DataNode: [[[DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest_A483062C73879C9B-001/tempDir-002/hdfsBaseDir/data/data4/]] heartbeating to localhost.localdomain/127.0.0.1:46029) [ ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
[junit4] 2> 1628972 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xa4871fd6f899b: from storage DS-ed7e664f-037a-4c90-9067-4d711de779af node DatanodeRegistration(127.0.0.1:45799, datanodeUuid=9a5392f1-a065-499d-a8b1-5fa36b32ee17, infoPort=37065, infoSecurePort=0, ipcPort=40531, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
[junit4] 2> 1628973 INFO (Block report processor) [ ] BlockStateChange BLOCK* processReport 0xa4871fd6f899b: from storage DS-d1098035-70ce-418b-964e-b4e17ae4ead3 node DatanodeRegistration(127.0.0.1:45799, datanodeUuid=9a5392f1-a065-499d-a8b1-5fa36b32ee17, infoPort=37065, infoSecurePort=0, ipcPort=40531, storageInfo=lv=-56;cid=testClusterID;nsid=453997229;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
[junit4] 2> 1629029 INFO (TEST-HdfsAutoAddReplicasIntegrationTest.testSimple-seed#[A483062C73879C9B]) [ ] o.a.s.SolrTestCaseJ4 ###Starting testSimple
[junit4] 2> 1629030 INFO (qtp1940206296-28728) [n:127.0.0.1:34223_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf&maxShardsPerNode=2&autoAddReplicas=true&name=testSimple1&nrtReplicas=2&action=CREATE&numShards=2&createNodeSet=127.0.0.1:45039_solr,127.0.0.1:39879_solr&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 1629032 INFO (OverseerThreadFactory-8464-thread-1) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection testSimple1
[junit4] 2> 1629135 INFO (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"testSimple1",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"testSimple1_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"http://127.0.0.1:39879/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1629136 INFO (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"testSimple1",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"testSimple1_shard1_replica_n2",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"http://127.0.0.1:45039/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1629137 INFO (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"testSimple1",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"testSimple1_shard2_replica_n4",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"http://127.0.0.1:39879/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1629138 INFO (OverseerStateUpdate-72247281650171911-127.0.0.1:45039_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"testSimple1",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"testSimple1_shard2_replica_n6",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"http://127.0.0.1:45039/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 1629290 DEBUG (ScheduledTrigger-8462-thread-3) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 3
[junit4] 2> 1629340 INFO (qtp1093226091-28721) [n:127.0.0.1:39879_solr x:testSimple1_shard2_replica_n4] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf&newCollection=true&name=testSimple1_shard2_replica_n4&action=CREATE&numShards=2&collection=testSimple1&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1629340 INFO (qtp347252057-28708) [n:127.0.0.1:45039_solr x:testSimple1_shard1_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf&newCollection=true&name=testSimple1_shard1_replica_n2&action=CREATE&numShards=2&collection=testSimple1&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1629341 INFO (qtp347252057-28708) [n:127.0.0.1:45039_solr x:testSimple1_shard1_replica_n2] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
[junit4] 2> 1629341 INFO (qtp347252057-28726) [n:127.0.0.1:45039_solr x:testSimple1_shard2_replica_n6] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf&newCollection=true&name=testSimple1_shard2_replica_n6&action=CREATE&numShards=2&collection=testSimple1&shard=shard2&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1629343 INFO (qtp1093226091-28711) [n:127.0.0.1:39879_solr x:testSimple1_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf&newCollection=true&name=testSimple1_shard1_replica_n1&action=CREATE&numShards=2&collection=testSimple1&shard=shard1&wt=javabin&version=2&replicaType=NRT
[junit4] 2> 1630291 DEBUG (ScheduledTrigger-8462-thread-4) [ ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 3
[junit4] 2> 1630346 INFO (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
[junit4] 2> 1630346 INFO (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
[junit4] 2> 1630360 INFO (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
[junit4] 2> 1630362 INFO (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
[junit4] 2> 1630363 INFO (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.s.IndexSchema [testSimple1_shard2_replica_n4] Schema name=minimal
[junit4] 2> 1630363 INFO (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.s.IndexSchema [testSimple1_shard2_replica_n6] Schema name=minimal
[junit4] 2> 1630365 INFO (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1630365 INFO (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard2_replica_n4' using configuration from collection testSimple1, trusted=true
[junit4] 2> 1630366 INFO (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1630366 INFO (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard2_replica_n6' using configuration from collection testSimple1, trusted=true
[junit4] 2> 1630366 INFO (qtp1093226091-28721) [n:127.0.0.1:39879_solr c:testSimple1 s:shard2 r:core_node7 x:testSimple1_shard2_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard2.replica_n4' (registry 'solr.core.testSimple1.shard2.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1630366 INFO (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard2.replica_n6' (registry 'solr.core.testSimple1.shard2.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1630367 INFO (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.s.IndexSchema [testSimple1_shard1_replica_n2] Schema name=minimal
[junit4] 2> 1630378 INFO (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.s.IndexSchema [testSimple1_shard1_replica_n1] Schema name=minimal
[junit4] 2> 1630378 INFO (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1630379 INFO (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard1_replica_n2' using configuration from collection testSimple1, trusted=true
[junit4] 2> 1630379 INFO (qtp347252057-28708) [n:127.0.0.1:45039_solr c:testSimple1 s:shard1 r:core_node5 x:testSimple1_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard1.replica_n2' (registry 'solr.core.testSimple1.shard1.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1630379 INFO (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 1630379 INFO (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'testSimple1_shard1_replica_n1' using configuration from collection testSimple1, trusted=true
[junit4] 2> 1630383 INFO (qtp1093226091-28711) [n:127.0.0.1:39879_solr c:testSimple1 s:shard1 r:core_node3 x:testSimple1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.testSimple1.shard1.replica_n1' (registry 'solr.core.testSimple1.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3086b156
[junit4] 2> 1630385 INFO (qtp347252057-28726) [n:127.0.0.1:45039_solr c:testSimple1 s:shard2 r:core_node8 x:testSimple1_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory solr.hdf
[...truncated too long message...]
build/contrib/solr-dataimporthandler/test/temp/junit4-J0-20180623_003429_74715312066844850799510.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-dataimporthandler/test/temp/junit4-J1-20180623_003429_7479827180275547199107.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 567 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J1-20180623_003457_7593588562413959039049.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 6 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J0-20180623_003457_7592461396733423052156.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-cell/test/temp/junit4-J2-20180623_003457_76214734783176774942234.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 576 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J2-20180623_003510_19510980844002221148259.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J0-20180623_003510_1958195709446057798390.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-langid/test/temp/junit4-J1-20180623_003510_1952916833743249538977.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 674 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J2-20180623_003520_44115365425148263749297.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J0-20180623_003520_44116354916187501290760.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 3 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-ltr/test/temp/junit4-J1-20180623_003520_44113703101826179663573.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 570 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-prometheus-exporter/test/temp/junit4-J0-20180623_003550_60611854339242176193155.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-prometheus-exporter/test/temp/junit4-J1-20180623_003550_6069919291497190540503.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 606 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-uima/test/temp/junit4-J1-20180623_003611_4283766516359282848020.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 3 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-uima/test/temp/junit4-J0-20180623_003611_42815424760474818802855.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 557 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/contrib/solr-velocity/test/temp/junit4-J0-20180623_003625_2359907790589491311819.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 34353 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22300/consoleText
[repro] Revision: 3b9d3a760a432b97aad2c08b2f778fa2344eb14a
[repro] Ant options: "-Dargs=-XX:-UseCompressedOops -XX:+UseConcMarkSweepGC"
[repro] Repro line: ant test -Dtestcase=HdfsAutoAddReplicasIntegrationTest -Dtests.method=testSimple -Dtests.seed=A483062C73879C9B -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ga -Dtests.timezone=Europe/Amsterdam -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1
[repro] ant clean
[...truncated 6 lines...]
[repro] Test suites by module:
[repro] solr/core
[repro] HdfsAutoAddReplicasIntegrationTest
[repro] ant compile-test
[...truncated 3305 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.HdfsAutoAddReplicasIntegrationTest" -Dtests.showOutput=onerror "-Dargs=-XX:-UseCompressedOops -XX:+UseConcMarkSweepGC" -Dtests.seed=A483062C73879C9B -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ga -Dtests.timezone=Europe/Amsterdam -Dtests.asserts=true -Dtests.file.encoding=ISO-8859-1
[...truncated 45 lines...]
[junit4] JVM J0: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J0-20180623_003921_9778145130587091982603.syserr
[junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J0: EOF ----
[...truncated 4 lines...]
[junit4] JVM J2: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J2-20180623_003921_97710617184633399344356.syserr
[junit4] >>> JVM J2 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J2: EOF ----
[...truncated 4 lines...]
[junit4] JVM J1: stderr was not empty, see: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/temp/junit4-J1-20180623_003921_97713048276681065257358.syserr
[junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
[junit4] OpenJDK 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
[junit4] <<< JVM J1: EOF ----
[...truncated 23 lines...]
[repro] Failures:
[repro] 0/5 failed: org.apache.solr.cloud.autoscaling.HdfsAutoAddReplicasIntegrationTest
[repro] Exiting with code 0
[...truncated 40 lines...]
[JENKINS] Lucene-Solr-master-Linux (64bit/jdk-10.0.1) - Build #
22301 - Still Unstable!
Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22301/
Java: 64bit/jdk-10.0.1 -XX:-UseCompressedOops -XX:+UseSerialGC
1 tests failed.
FAILED: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest.testBiDir
Error Message:
Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
Stack Trace:
com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
Caused by: java.lang.AssertionError
at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
at java.base/java.lang.Thread.run(Thread.java:844)
Build Log:
[...truncated 13442 lines...]
[junit4] Suite: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest
[junit4] 2> 726810 INFO (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [ ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
[junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/init-core-data-001
[junit4] 2> 726813 WARN (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [ ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=25 numCloses=25
[junit4] 2> 726813 INFO (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [ ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
[junit4] 2> 726814 INFO (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [ ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (true) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0)
[junit4] 2> 726821 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.SolrTestCaseJ4 ###Starting testBiDir
[junit4] 2> 726821 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 1 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001
[junit4] 2> 726822 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 726834 INFO (Thread-2493) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 726834 INFO (Thread-2493) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 726867 ERROR (Thread-2493) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 726934 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.ZkTestServer start zk server on port:38873
[junit4] 2> 726967 INFO (zkConnectionManagerCallback-2451-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 726987 INFO (jetty-launcher-2448-thread-1) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 727006 INFO (jetty-launcher-2448-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 727006 INFO (jetty-launcher-2448-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 727006 INFO (jetty-launcher-2448-thread-1) [ ] o.e.j.s.session node0 Scavenging every 600000ms
[junit4] 2> 727007 INFO (jetty-launcher-2448-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@7c514370{/solr,null,AVAILABLE}
[junit4] 2> 727008 INFO (jetty-launcher-2448-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@55d7ae79{SSL,[ssl, http/1.1]}{127.0.0.1:৩৯৩৪৯}
[junit4] 2> 727008 INFO (jetty-launcher-2448-thread-1) [ ] o.e.j.s.Server Started @à§à§¨à§à§¦à§©à§©ms
[junit4] 2> 727008 INFO (jetty-launcher-2448-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=39349}
[junit4] 2> 727008 ERROR (jetty-launcher-2448-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 727008 INFO (jetty-launcher-2448-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 727008 INFO (jetty-launcher-2448-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solrâ„¢ version 8.0.0
[junit4] 2> 727008 INFO (jetty-launcher-2448-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 727008 INFO (jetty-launcher-2448-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 727008 INFO (jetty-launcher-2448-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-23T04:08:53.265596Z
[junit4] 2> 727034 INFO (zkConnectionManagerCallback-2453-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727035 INFO (jetty-launcher-2448-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 727144 INFO (jetty-launcher-2448-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:38873/solr
[junit4] 2> 727145 INFO (zkConnectionManagerCallback-2457-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727146 INFO (zkConnectionManagerCallback-2459-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727181 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 727182 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:39349_solr
[junit4] 2> 727182 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.c.Overseer Overseer (id=72248175661154307-127.0.0.1:39349_solr-n_0000000000) starting
[junit4] 2> 727186 INFO (zkConnectionManagerCallback-2466-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727188 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
[junit4] 2> 727188 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:39349_solr
[junit4] 2> 727188 INFO (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 727198 INFO (zkCallback-2465-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 727210 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 727225 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 727232 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 727232 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 727233 INFO (jetty-launcher-2448-thread-1) [n:127.0.0.1:39349_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/.
[junit4] 2> 727254 INFO (zkConnectionManagerCallback-2469-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727258 INFO (zkConnectionManagerCallback-2472-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727259 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 1 servers in /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001
[junit4] 2> 727260 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
[junit4] 2> 727260 INFO (Thread-2503) [ ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 727260 INFO (Thread-2503) [ ] o.a.s.c.ZkTestServer Starting server
[junit4] 2> 727267 ERROR (Thread-2503) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 727360 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.ZkTestServer start zk server on port:44483
[junit4] 2> 727361 INFO (zkConnectionManagerCallback-2476-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727371 INFO (jetty-launcher-2473-thread-1) [ ] o.e.j.s.Server jetty-9.4.10.v20180503; built: 2018-05-03T15:56:21.710Z; git: daa59876e6f384329b122929e70a80934569428c; jvm 10.0.1+10
[junit4] 2> 727373 INFO (jetty-launcher-2473-thread-1) [ ] o.e.j.s.session DefaultSessionIdManager workerName=node0
[junit4] 2> 727373 INFO (jetty-launcher-2473-thread-1) [ ] o.e.j.s.session No SessionScavenger set, using defaults
[junit4] 2> 727373 INFO (jetty-launcher-2473-thread-1) [ ] o.e.j.s.session node0 Scavenging every 660000ms
[junit4] 2> 727374 INFO (jetty-launcher-2473-thread-1) [ ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@6ad5a77d{/solr,null,AVAILABLE}
[junit4] 2> 727374 INFO (jetty-launcher-2473-thread-1) [ ] o.e.j.s.AbstractConnector Started ServerConnector@3cfe7871{SSL,[ssl, http/1.1]}{127.0.0.1:৪৩à§à§¯à§§}
[junit4] 2> 727374 INFO (jetty-launcher-2473-thread-1) [ ] o.e.j.s.Server Started @à§à§¨à§à§ªà§¦à§¦ms
[junit4] 2> 727374 INFO (jetty-launcher-2473-thread-1) [ ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=43791}
[junit4] 2> 727378 ERROR (jetty-launcher-2473-thread-1) [ ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
[junit4] 2> 727378 INFO (jetty-launcher-2473-thread-1) [ ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
[junit4] 2> 727378 INFO (jetty-launcher-2473-thread-1) [ ] o.a.s.s.SolrDispatchFilter ___ _ Welcome to Apache Solrâ„¢ version 8.0.0
[junit4] 2> 727378 INFO (jetty-launcher-2473-thread-1) [ ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _ Starting in cloud mode on port null
[junit4] 2> 727378 INFO (jetty-launcher-2473-thread-1) [ ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_| Install dir: null
[junit4] 2> 727378 INFO (jetty-launcher-2473-thread-1) [ ] o.a.s.s.SolrDispatchFilter |___/\___/_|_| Start time: 2018-06-23T04:08:53.635561Z
[junit4] 2> 727380 INFO (zkConnectionManagerCallback-2478-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727381 INFO (jetty-launcher-2473-thread-1) [ ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
[junit4] 2> 727532 INFO (jetty-launcher-2473-thread-1) [ ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44483/solr
[junit4] 2> 727543 INFO (zkConnectionManagerCallback-2482-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727546 INFO (zkConnectionManagerCallback-2484-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727582 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.c.Overseer Overseer (id=null) closing
[junit4] 2> 727582 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:43791_solr
[junit4] 2> 727583 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.c.Overseer Overseer (id=72248175688417283-127.0.0.1:43791_solr-n_0000000000) starting
[junit4] 2> 727590 INFO (zkConnectionManagerCallback-2491-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727592 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44483/solr ready
[junit4] 2> 727601 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:43791_solr
[junit4] 2> 727618 INFO (zkCallback-2483-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 727623 INFO (zkCallback-2490-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 727632 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
[junit4] 2> 727641 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 727646 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 727646 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 727647 INFO (jetty-launcher-2473-thread-1) [n:127.0.0.1:43791_solr ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/.
[junit4] 2> 727674 INFO (zkConnectionManagerCallback-2494-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727698 INFO (zkConnectionManagerCallback-2497-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727703 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.c.CdcrBidirectionalTest cluster2 zkHost = 127.0.0.1:38873/solr
[junit4] 2> 727703 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.c.CdcrBidirectionalTest cluster1 zkHost = 127.0.0.1:44483/solr
[junit4] 2> 727710 INFO (zkConnectionManagerCallback-2499-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727722 INFO (zkConnectionManagerCallback-2503-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 727723 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 727724 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44483/solr ready
[junit4] 2> 727738 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=cdcr-cluster1&maxShardsPerNode=2&name=cdcr-cluster1&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 727742 INFO (OverseerThreadFactory-3404-thread-1) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection cdcr-cluster1
[junit4] 2> 727847 INFO (OverseerStateUpdate-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"cdcr-cluster1",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"cdcr-cluster1_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:43791/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 727850 INFO (OverseerStateUpdate-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"cdcr-cluster1",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"cdcr-cluster1_shard2_replica_n3",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:43791/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 728103 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster1_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
[junit4] 2> 728104 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 728103 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster1_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
[junit4] 2> 729144 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
[junit4] 2> 729149 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.IndexSchema [cdcr-cluster1_shard1_replica_n1] Schema name=minimal
[junit4] 2> 729150 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 729150 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster1_shard1_replica_n1' using configuration from collection cdcr-cluster1, trusted=true
[junit4] 2> 729151 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
[junit4] 2> 729152 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.core.cdcr-cluster1.shard1.replica_n1' (registry 'solr.core.cdcr-cluster1.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 729154 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 729156 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [[cdcr-cluster1_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/cdcr-cluster1_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/./cdcr-cluster1_shard1_replica_n1/data/]
[junit4] 2> 729156 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.IndexSchema [cdcr-cluster1_shard2_replica_n3] Schema name=minimal
[junit4] 2> 729157 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 729157 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster1_shard2_replica_n3' using configuration from collection cdcr-cluster1, trusted=true
[junit4] 2> 729158 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_43791.solr.core.cdcr-cluster1.shard2.replica_n3' (registry 'solr.core.cdcr-cluster1.shard2.replica_n3') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 729158 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 729158 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [[cdcr-cluster1_shard2_replica_n3] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/cdcr-cluster1_shard2_replica_n3], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster1-001/node1/./cdcr-cluster1_shard2_replica_n3/data/]
[junit4] 2> 729331 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
[junit4] 2> 729331 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
[junit4] 2> 729331 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 729331 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 729332 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 729332 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 729332 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 729332 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 729333 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@4cade66a[cdcr-cluster1_shard1_replica_n1] main]
[junit4] 2> 729333 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@534a2f15[cdcr-cluster1_shard2_replica_n3] main]
[junit4] 2> 729334 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster1
[junit4] 2> 729334 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster1
[junit4] 2> 729334 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster1
[junit4] 2> 729334 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster1
[junit4] 2> 729334 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster1/managed-schema
[junit4] 2> 729334 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster1/managed-schema
[junit4] 2> 729334 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
[junit4] 2> 729334 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
[junit4] 2> 729335 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 729335 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 729338 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.h.CdcrBufferStateManager Created znode /collections/cdcr-cluster1/cdcr/state/buffer
[junit4] 2> 729340 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrProcessStateManager Created znode /collections/cdcr-cluster1/cdcr/state/process
[junit4] 2> 729353 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034951234191360
[junit4] 2> 729353 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034951234191360
[junit4] 2> 729354 INFO (searcherExecutor-3410-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard2_replica_n3 c:cdcr-cluster1 s:shard2 r:core_node4) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster1_shard2_replica_n3] Registered new searcher Searcher@534a2f15[cdcr-cluster1_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 729354 INFO (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@4cade66a[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 729357 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard1 to Terms{values={core_node2=0}, version=0}
[junit4] 2> 729358 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard2 to Terms{values={core_node4=0}, version=0}
[junit4] 2> 729359 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 729359 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 729359 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/
[junit4] 2> 729359 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 729359 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/ has no replicas
[junit4] 2> 729359 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 729360 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 729360 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 729360 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/
[junit4] 2> 729360 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 729360 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SyncStrategy https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/ has no replicas
[junit4] 2> 729360 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 729361 INFO (zkCallback-2483-thread-1) [ ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster1:shard1
[junit4] 2> 729364 INFO (zkCallback-2483-thread-1) [ ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster1:shard2
[junit4] 2> 729364 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/ shard1
[junit4] 2> 729366 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/ shard2
[junit4] 2> 729476 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 729477 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster1_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1374
[junit4] 2> 729516 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 729517 INFO (qtp804107552-10557) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster1&newCollection=true&collection=cdcr-cluster1&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster1_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1414
[junit4] 2> 729519 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
[junit4] 2> 729618 INFO (zkCallback-2483-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/state.json] for collection [cdcr-cluster1] has occurred - updating... (live nodes size: [1])
[junit4] 2> 729744 INFO (OverseerCollectionConfigSetProcessor-72248175688417283-127.0.0.1:43791_solr-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 730519 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=cdcr-cluster1&maxShardsPerNode=2&name=cdcr-cluster1&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2} status=0 QTime=2781
[junit4] 2> 730538 INFO (zkConnectionManagerCallback-2507-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 730545 INFO (zkConnectionManagerCallback-2511-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 730546 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 730560 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
[junit4] 2> 730623 INFO (qtp185556678-10502) [n:127.0.0.1:39349_solr ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=cdcr-cluster2&maxShardsPerNode=2&name=cdcr-cluster2&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2 and sendToOCPQueue=true
[junit4] 2> 730639 INFO (OverseerThreadFactory-3391-thread-1) [ ] o.a.s.c.a.c.CreateCollectionCmd Create collection cdcr-cluster2
[junit4] 2> 730749 INFO (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"cdcr-cluster2",
[junit4] 2> "shard":"shard1",
[junit4] 2> "core":"cdcr-cluster2_shard1_replica_n1",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:39349/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 730751 INFO (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [ ] o.a.s.c.o.SliceMutator createReplica() {
[junit4] 2> "operation":"ADDREPLICA",
[junit4] 2> "collection":"cdcr-cluster2",
[junit4] 2> "shard":"shard2",
[junit4] 2> "core":"cdcr-cluster2_shard2_replica_n3",
[junit4] 2> "state":"down",
[junit4] 2> "base_url":"https://127.0.0.1:39349/solr",
[junit4] 2> "type":"NRT",
[junit4] 2> "waitForFinalState":"false"}
[junit4] 2> 730989 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster2_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
[junit4] 2> 730990 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
[junit4] 2> 731031 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster2_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin
[junit4] 2> 732008 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
[junit4] 2> 732013 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.IndexSchema [cdcr-cluster2_shard1_replica_n1] Schema name=minimal
[junit4] 2> 732043 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 732044 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster2_shard1_replica_n1' using configuration from collection cdcr-cluster2, trusted=true
[junit4] 2> 732044 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.core.cdcr-cluster2.shard1.replica_n1' (registry 'solr.core.cdcr-cluster2.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 732044 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 732044 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [[cdcr-cluster2_shard1_replica_n1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/cdcr-cluster2_shard1_replica_n1], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/./cdcr-cluster2_shard1_replica_n1/data/]
[junit4] 2> 732055 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrConfig Using Lucene MatchVersion: 8.0.0
[junit4] 2> 732063 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.IndexSchema [cdcr-cluster2_shard2_replica_n3] Schema name=minimal
[junit4] 2> 732064 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
[junit4] 2> 732064 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.CoreContainer Creating SolrCore 'cdcr-cluster2_shard2_replica_n3' using configuration from collection cdcr-cluster2, trusted=true
[junit4] 2> 732065 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_39349.solr.core.cdcr-cluster2.shard2.replica_n3' (registry 'solr.core.cdcr-cluster2.shard2.replica_n3') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@6218a436
[junit4] 2> 732065 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
[junit4] 2> 732065 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [[cdcr-cluster2_shard2_replica_n3] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/cdcr-cluster2_shard2_replica_n3], dataDir=[/home/jenkins/workspace/Lucene-Solr-master-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.cdcr.CdcrBidirectionalTest_27F8202107443866-001/cdcr-cluster2-001/node1/./cdcr-cluster2_shard2_replica_n3/data/]
[junit4] 2> 732150 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
[junit4] 2> 732151 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 732152 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 732152 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 732155 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@116811f2[cdcr-cluster2_shard1_replica_n1] main]
[junit4] 2> 732156 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster2
[junit4] 2> 732156 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster2
[junit4] 2> 732156 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster2/managed-schema
[junit4] 2> 732156 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
[junit4] 2> 732157 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 732159 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrBufferStateManager Created znode /collections/cdcr-cluster2/cdcr/state/buffer
[junit4] 2> 732165 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrProcessStateManager Created znode /collections/cdcr-cluster2/cdcr/state/process
[junit4] 2> 732206 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034954225778688
[junit4] 2> 732208 INFO (searcherExecutor-3421-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster2_shard1_replica_n1] Registered new searcher Searcher@116811f2[cdcr-cluster2_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 732208 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.CdcrUpdateLog
[junit4] 2> 732208 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
[junit4] 2> 732209 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.CommitTracker Hard AutoCommit: disabled
[junit4] 2> 732209 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.CommitTracker Soft AutoCommit: disabled
[junit4] 2> 732211 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster2/terms/shard1 to Terms{values={core_node2=0}, version=0}
[junit4] 2> 732212 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@7910e119[cdcr-cluster2_shard2_replica_n3] main]
[junit4] 2> 732212 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/cdcr-cluster2
[junit4] 2> 732213 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/cdcr-cluster2
[junit4] 2> 732213 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/cdcr-cluster2/managed-schema
[junit4] 2> 732213 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
[junit4] 2> 732214 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
[junit4] 2> 732215 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 732215 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 732215 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/
[junit4] 2> 732215 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 732215 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SyncStrategy https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/ has no replicas
[junit4] 2> 732215 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 732216 INFO (zkCallback-2458-thread-2) [ ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster2:shard1
[junit4] 2> 732217 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/ shard1
[junit4] 2> 732224 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1604034954244653056
[junit4] 2> 732252 INFO (searcherExecutor-3422-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster2_shard2_replica_n3] Registered new searcher Searcher@7910e119[cdcr-cluster2_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 732253 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster2/terms/shard2 to Terms{values={core_node4=0}, version=0}
[junit4] 2> 732255 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
[junit4] 2> 732255 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
[junit4] 2> 732255 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy Sync replicas to https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/
[junit4] 2> 732255 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
[junit4] 2> 732255 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SyncStrategy https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/ has no replicas
[junit4] 2> 732255 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
[junit4] 2> 732256 INFO (zkCallback-2458-thread-2) [ ] o.a.s.h.CdcrLeaderStateManager Received new leader state @ cdcr-cluster2:shard2
[junit4] 2> 732258 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ShardLeaderElectionContext I am the new leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/ shard2
[junit4] 2> 732360 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 732361 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node4&name=cdcr-cluster2_shard2_replica_n3&action=CREATE&numShards=2&shard=shard2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1330
[junit4] 2> 732368 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
[junit4] 2> 732369 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=cdcr-cluster2&newCollection=true&collection=cdcr-cluster2&version=2&replicaType=NRT&coreNodeName=core_node2&name=cdcr-cluster2_shard1_replica_n1&action=CREATE&numShards=2&shard=shard1&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin} status=0 QTime=1380
[junit4] 2> 732373 INFO (qtp185556678-10502) [n:127.0.0.1:39349_solr ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
[junit4] 2> 732470 INFO (zkCallback-2458-thread-1) [ ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster2/state.json] for collection [cdcr-cluster2] has occurred - updating... (live nodes size: [1])
[junit4] 2> 732662 INFO (OverseerCollectionConfigSetProcessor-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [ ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist. Requestor may have disconnected from ZooKeeper
[junit4] 2> 733373 INFO (qtp185556678-10502) [n:127.0.0.1:39349_solr ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=cdcr-cluster2&maxShardsPerNode=2&name=cdcr-cluster2&nrtReplicas=1&action=CREATE&numShards=2&property.solr.directoryFactory=solr.StandardDirectoryFactory&wt=javabin&version=2} status=0 QTime=2750
[junit4] 2> 733390 INFO (zkConnectionManagerCallback-2518-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 733391 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 733391 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
[junit4] 2> 733421 INFO (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
[junit4] 2> 733421 INFO (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=3
[junit4] 2> 733443 INFO (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
[junit4] 2> 733443 INFO (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1] webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 733444 INFO (qtp185556678-10496) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=45
[junit4] 2> 733445 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard1
[junit4] 2> 733445 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2, shard: shard1
[junit4] 2> 733447 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Submitting bootstrap task to executor
[junit4] 2> 733451 INFO (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2 shard: shard1 leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard1_replica_n1/
[junit4] 2> 733461 INFO (zkCallback-2483-thread-1) [ ] o.a.s.h.CdcrProcessStateManager The CDCR process state has changed: WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/cdcr/state/process @ cdcr-cluster1:shard2
[junit4] 2> 733461 INFO (zkCallback-2483-thread-2) [ ] o.a.s.h.CdcrProcessStateManager The CDCR process state has changed: WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/cdcr-cluster1/cdcr/state/process @ cdcr-cluster1:shard1
[junit4] 2> 733462 INFO (zkCallback-2483-thread-1) [ ] o.a.s.h.CdcrProcessStateManager Received new CDCR process state from watcher: STARTED @ cdcr-cluster1:shard2
[junit4] 2> 733462 INFO (zkCallback-2483-thread-2) [ ] o.a.s.h.CdcrProcessStateManager Received new CDCR process state from watcher: STARTED @ cdcr-cluster1:shard1
[junit4] 2> 733462 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1] webapp=/solr path=/cdcr params={qt=/cdcr&_stateVer_=cdcr-cluster1:4&action=start&wt=javabin&version=2} status=0 QTime=85
[junit4] 2> 733514 INFO (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1] webapp=/solr path=/cdcr params={qt=/cdcr&masterUrl=https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/&action=BOOTSTRAP&wt=javabin&version=2} status=0 QTime=5
[junit4] 2> 733516 INFO (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1] webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 733518 INFO (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager CDCR bootstrap running for 1 seconds, sleeping for 2000 ms
[junit4] 2> 733526 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
[junit4] 2> 733534 INFO (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034955618287616,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 733542 INFO (zkConnectionManagerCallback-2523-thread-1) [ ] o.a.s.c.c.ConnectionManager zkClient has connected
[junit4] 2> 733542 INFO (zkCallback-2483-thread-1) [ ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
[junit4] 2> 733543 INFO (zkCallback-2483-thread-1) [ ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:38873/solr ready
[junit4] 2> 733554 INFO (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
[junit4] 2> 733555 INFO (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 733555 INFO (qtp804107552-10550) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1] webapp=/solr path=/update params={waitSearcher=true&openSearcher=false&commit=true&softCommit=false&commit_end_point=true&wt=javabin&version=2} status=0 QTime=22
[junit4] 2> 733585 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
[junit4] 2> 733585 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 733609 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
[junit4] 2> 733609 INFO (qtp185556678-10503) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1] webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 733610 INFO (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=64
[junit4] 2> 733610 INFO (zkCallback-2483-thread-1) [ ] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard2
[junit4] 2> 733610 INFO (zkCallback-2483-thread-1) [ ] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2, shard: shard2
[junit4] 2> 733610 INFO (zkCallback-2483-thread-1) [ ] o.a.s.h.CdcrReplicatorManager Submitting bootstrap task to executor
[junit4] 2> 733615 INFO (cdcr-bootstrap-status-2519-thread-1) [ ] o.a.s.h.CdcrReplicatorManager Attempting to bootstrap target collection: cdcr-cluster2 shard: shard2 leader: https://127.0.0.1:39349/solr/cdcr-cluster2_shard2_replica_n3/
[junit4] 2> 733622 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1] webapp=/solr path=/replication params={qt=/replication&wt=javabin&version=2&command=indexversion} status=0 QTime=0
[junit4] 2> 733622 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Master's generation: 1
[junit4] 2> 733622 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Master's version: 0
[junit4] 2> 733622 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Slave's generation: 1
[junit4] 2> 733622 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher Slave's version: 0
[junit4] 2> 733622 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.IndexFetcher New index in Master. Deleting mine...
[junit4] 2> 733627 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@860e11[cdcr-cluster2_shard1_replica_n1] main]
[junit4] 2> 733628 INFO (searcherExecutor-3421-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster2_shard1_replica_n1] Registered new searcher Searcher@860e11[cdcr-cluster2_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 733628 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard1_replica_n1 c:cdcr-cluster2 s:shard1 r:core_node2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler No replay needed.
[junit4] 2> 733628 INFO (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={qt=/cdcr&masterUrl=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&action=BOOTSTRAP&wt=javabin&version=2} status=0 QTime=9
[junit4] 2> 733629 INFO (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 733630 INFO (cdcr-bootstrap-status-2519-thread-1) [ ] o.a.s.h.CdcrReplicatorManager CDCR bootstrap running for 1 seconds, sleeping for 2000 ms
[junit4] 2> 733630 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.u.UpdateLog Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
[junit4] 2> 733641 INFO (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034955730485248,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 733646 INFO (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
[junit4] 2> 733646 INFO (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 733646 INFO (qtp804107552-10552) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3] webapp=/solr path=/update params={waitSearcher=true&openSearcher=false&commit=true&softCommit=false&commit_end_point=true&wt=javabin&version=2} status=0 QTime=6
[junit4] 2> 733650 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3] webapp=/solr path=/replication params={qt=/replication&wt=javabin&version=2&command=indexversion} status=0 QTime=0
[junit4] 2> 733650 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Master's generation: 1
[junit4] 2> 733650 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Master's version: 0
[junit4] 2> 733650 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Slave's generation: 1
[junit4] 2> 733650 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher Slave's version: 0
[junit4] 2> 733650 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.IndexFetcher New index in Master. Deleting mine...
[junit4] 2> 733651 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@3853640e[cdcr-cluster2_shard2_replica_n3] main]
[junit4] 2> 733652 INFO (searcherExecutor-3422-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster2_shard2_replica_n3] Registered new searcher Searcher@3853640e[cdcr-cluster2_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 733652 INFO (recoveryExecutor-2455-thread-1-processing-n:127.0.0.1:39349_solr x:cdcr-cluster2_shard2_replica_n3 c:cdcr-cluster2 s:shard2 r:core_node4) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler No replay needed.
[junit4] 2> 735463 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.c.CdcrBidirectionalTest Adding 10 docs with commit=true, numDocs=100
[junit4] 2> 735469 INFO (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard1 to Terms{values={core_node2=1}, version=1}
[junit4] 2> 735469 INFO (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1] webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=3
[junit4] 2> 735469 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.ZkShardTerms Successful update of terms at /collections/cdcr-cluster1/terms/shard2 to Terms{values={core_node4=1}, version=1}
[junit4] 2> 735469 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3] webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=5
[junit4] 2> 735474 INFO (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957652525056,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 735474 INFO (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@62bc2cb1 commitCommandVersion:1604034957652525056
[junit4] 2> 735478 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957656719360,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 735478 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@6850d084 commitCommandVersion:1604034957656719360
[junit4] 2> 735518 INFO (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1] webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 735519 INFO (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager CDCR bootstrap successful in 3 seconds
[junit4] 2> 735523 INFO (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
[junit4] 2> 735523 INFO (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1] webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 735525 INFO (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
[junit4] 2> 735525 INFO (qtp185556678-10499) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 735528 INFO (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=9
[junit4] 2> 735529 INFO (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard1
[junit4] 2> 735529 INFO (cdcr-bootstrap-status-2514-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.h.CdcrReplicatorManager Bootstrap successful, giving the go-ahead to replicator
[junit4] 2> 735630 INFO (qtp185556678-10628) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={qt=/cdcr&action=BOOTSTRAP_STATUS&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 735631 INFO (cdcr-bootstrap-status-2519-thread-1) [ ] o.a.s.h.CdcrReplicatorManager CDCR bootstrap successful in 3 seconds
[junit4] 2> 735637 INFO (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
[junit4] 2> 735637 INFO (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.h.CdcrRequestHandler Found maxVersionFromRecent 0 maxVersionFromIndex 0
[junit4] 2> 735637 INFO (qtp185556678-10497) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 735637 INFO (qtp185556678-10501) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard1 r:core_node2 x:cdcr-cluster2_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster2_shard1_replica_n1] webapp=/solr path=/cdcr params={action=SHARDCHECKPOINT&wt=javabin&version=2} status=0 QTime=0
[junit4] 2> 735638 INFO (qtp185556678-10502) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster2_shard2_replica_n3] webapp=/solr path=/cdcr params={_stateVer_=cdcr-cluster2:4&action=COLLECTIONCHECKPOINT&wt=javabin&version=2} status=0 QTime=6
[junit4] 2> 735638 INFO (cdcr-bootstrap-status-2519-thread-1) [ ] o.a.s.h.CdcrReplicatorManager Create new update log reader for target cdcr-cluster2 with checkpoint -1 @ cdcr-cluster1:shard2
[junit4] 2> 735638 INFO (cdcr-bootstrap-status-2519-thread-1) [ ] o.a.s.h.CdcrReplicatorManager Bootstrap successful, giving the go-ahead to replicator
[junit4] 2> 735653 INFO (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@601f19c[cdcr-cluster1_shard1_replica_n1] main]
[junit4] 2> 735654 INFO (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 735656 INFO (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@601f19c[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C43)))}
[junit4] 2> 735656 INFO (qtp804107552-10608) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1] webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false} status=0 QTime=182
[junit4] 2> 735667 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.s.SolrIndexSearcher Opening [Searcher@6fb4ca9f[cdcr-cluster1_shard2_replica_n3] main]
[junit4] 2> 735668 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 735669 INFO (searcherExecutor-3410-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard2_replica_n3 c:cdcr-cluster1 s:shard2 r:core_node4) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.SolrCore [cdcr-cluster1_shard2_replica_n3] Registered new searcher Searcher@6fb4ca9f[cdcr-cluster1_shard2_replica_n3] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C57)))}
[junit4] 2> 735669 INFO (qtp804107552-10555) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3] webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard2_replica_n3/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false} status=0 QTime=191
[junit4] 2> 735669 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3] webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&waitSearcher=true&commit=true&softCommit=false&wt=javabin&version=2} status=0 QTime=199
[junit4] 2> 735670 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.c.CdcrBidirectionalTest Adding 10 docs with commit=true, numDocs=200
[junit4] 2> 735680 INFO (qtp804107552-10610) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1] webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=9
[junit4] 2> 735683 INFO (qtp804107552-10609) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.c.S.Request [cdcr-cluster1_shard2_replica_n3] webapp=/solr path=/update params={_stateVer_=cdcr-cluster1:4&wt=javabin&version=2} status=0 QTime=6
[junit4] 2> 735684 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957872726016,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 735684 INFO (qtp804107552-10553) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard2 r:core_node4 x:cdcr-cluster1_shard2_replica_n3] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@6850d084 commitCommandVersion:1604034957872726016
[junit4] 2> 735686 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1604034957874823168,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
[junit4] 2> 735686 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.SolrIndexWriter Calling setCommitData with IW:org.apache.solr.update.SolrIndexWriter@62bc2cb1 commitCommandVersion:1604034957874823168
[junit4] 2> 735972 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@3f3de49e[cdcr-cluster1_shard1_replica_n1] main]
[junit4] 2> 735973 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
[junit4] 2> 735974 INFO (searcherExecutor-3409-thread-1-processing-n:127.0.0.1:43791_solr x:cdcr-cluster1_shard1_replica_n1 c:cdcr-cluster1 s:shard1 r:core_node2) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.SolrCore [cdcr-cluster1_shard1_replica_n1] Registered new searcher Searcher@3f3de49e[cdcr-cluster1_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader(Uninverting(_0(8.0.0):C43) Uninverting(_1(8.0.0):C54)))}
[junit4] 2> 735974 INFO (qtp804107552-10556) [n:127.0.0.1:43791_solr c:cdcr-cluster1 s:shard1 r:core_node2 x:cdcr-cluster1_shard1_replica_n1] o.a.s.c.S.Request [cdcr-cluster1_shard1_replica_n1] webapp=/solr path=/update params={update.distrib=FROMLEADER&update.chain=cdcr-processor-chain&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=https://127.0.0.1:43791/solr/cdcr-cluster1_shard1_replica_n1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}
[...truncated too long message...]
try=solr.collection.cdcr-cluster2.shard2.leader, tag=18455e69
[junit4] 2> 760651 INFO (coreCloseExecutor-3451-thread-2) [n:127.0.0.1:39349_solr c:cdcr-cluster2 s:shard2 r:core_node4 x:cdcr-cluster2_shard2_replica_n3] o.a.s.h.CdcrRequestHandler Solr core is being closed - shutting down CDCR handler @ cdcr-cluster2:shard2
[junit4] 2> 760731 INFO (jetty-closer-2449-thread-1) [ ] o.a.s.c.Overseer Overseer (id=72248175661154307-127.0.0.1:39349_solr-n_0000000000) closing
[junit4] 2> 760731 INFO (OverseerStateUpdate-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [ ] o.a.s.c.Overseer Overseer Loop exiting : 127.0.0.1:39349_solr
[junit4] 2> 760731 WARN (OverseerAutoScalingTriggerThread-72248175661154307-127.0.0.1:39349_solr-n_0000000000) [ ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread woken up but we are closed, exiting.
[junit4] 2> 760732 INFO (jetty-closer-2449-thread-1) [ ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@7c514370{/solr,null,UNAVAILABLE}
[junit4] 2> 760732 INFO (jetty-closer-2449-thread-1) [ ] o.e.j.s.session node0 Stopped scavenging
[junit4] 2> 760733 ERROR (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
[junit4] 2> 760733 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:38873 38873
[junit4] 2> 760778 INFO (Thread-2493) [ ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:38873 38873
[junit4] 2> 760778 WARN (Thread-2493) [ ] o.a.s.c.ZkTestServer Watch limit violations:
[junit4] 2> Maximum concurrent create/delete watches above limit:
[junit4] 2>
[junit4] 2> 5 /solr/aliases.json
[junit4] 2> 3 /solr/collections/cdcr-cluster2/terms/shard2
[junit4] 2> 3 /solr/collections/cdcr-cluster2/terms/shard1
[junit4] 2> 2 /solr/configs/cdcr-cluster2/managed-schema
[junit4] 2>
[junit4] 2> Maximum concurrent data watches above limit:
[junit4] 2>
[junit4] 2> 6 /solr/collections/cdcr-cluster2/cdcr/state/process
[junit4] 2> 5 /solr/clusterstate.json
[junit4] 2> 5 /solr/clusterprops.json
[junit4] 2> 3 /solr/collections/cdcr-cluster2/state.json
[junit4] 2> 2 /solr/collections/cdcr-cluster2/cdcr/state/buffer
[junit4] 2>
[junit4] 2> Maximum concurrent children watches above limit:
[junit4] 2>
[junit4] 2> 5 /solr/live_nodes
[junit4] 2> 5 /solr/collections
[junit4] 2>
[junit4] 2> 760779 INFO (TEST-CdcrBidirectionalTest.testBiDir-seed#[27F8202107443866]) [ ] o.a.s.SolrTestCaseJ4 ###Ending testBiDir
[junit4] 2> NOTE: reproduce with: ant test -Dtestcase=CdcrBidirectionalTest -Dtests.method=testBiDir -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[junit4] ERROR 34.0s J0 | CdcrBidirectionalTest.testBiDir <<<
[junit4] > Throwable #1: com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10709, name=cdcr-replicator-3446-thread-1, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
[junit4] > Caused by: java.lang.AssertionError
[junit4] > at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
[junit4] > at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
[junit4] > at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
[junit4] > at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
[junit4] > at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
[junit4] > at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
[junit4] > at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
[junit4] > at java.base/java.lang.Thread.run(Thread.java:844)Throwable #2: com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=10729, name=cdcr-replicator-3446-thread-2, state=RUNNABLE, group=TGRP-CdcrBidirectionalTest]
[junit4] > Caused by: java.lang.AssertionError
[junit4] > at __randomizedtesting.SeedInfo.seed([27F8202107443866]:0)
[junit4] > at org.apache.solr.update.CdcrUpdateLog$CdcrLogReader.forwardSeek(CdcrUpdateLog.java:611)
[junit4] > at org.apache.solr.handler.CdcrReplicator.run(CdcrReplicator.java:125)
[junit4] > at org.apache.solr.handler.CdcrReplicatorScheduler.lambda$start$0(CdcrReplicatorScheduler.java:81)
[junit4] > at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209)
[junit4] > at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1135)
[junit4] > at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
[junit4] > at java.base/java.lang.Thread.run(Thread.java:844)
[junit4] 2> 760780 INFO (SUITE-CdcrBidirectionalTest-seed#[27F8202107443866]-worker) [ ] o.a.s.SolrTestCaseJ4 ------------------------------------------------------- Done waiting for tracked resources to be released
[junit4] 2> Jun 23, 2018 4:09:27 AM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
[junit4] 2> WARNING: Will linger awaiting termination of 5 leaked thread(s).
[junit4] 2> NOTE: test params are: codec=Asserting(Lucene70), sim=Asserting(org.apache.lucene.search.similarities.AssertingSimilarity@8d81519), locale=bn-IN, timezone=Asia/Tomsk
[junit4] 2> NOTE: Linux 4.13.0-41-generic amd64/Oracle Corporation 10.0.1 (64-bit)/cpus=8,threads=1,free=149515200,total=518979584
[junit4] 2> NOTE: All tests run in this JVM: [TestNestedDocsSort, TestMaxScoreQueryParser, URLClassifyProcessorTest, ZkCLITest, TestHttpShardHandlerFactory, SimpleCollectionCreateDeleteTest, TestClusterStateMutator, TestMinMaxOnMultiValuedField, ZkNodePropsTest, TestSort, TestComplexPhraseLeadingWildcard, TestHdfsBackupRestoreCore, TestSystemIdResolver, CursorPagingTest, TestSolrCloudWithDelegationTokens, ChaosMonkeyNothingIsSafeTest, TestBulkSchemaConcurrent, TestCollationFieldDocValues, DeleteLastCustomShardedReplicaTest, TestNumericRangeQuery32, TestRTGBase, TestManagedSchemaAPI, FileUtilsTest, BlockJoinFacetSimpleTest, TestShortCircuitedRequests, TestJettySolrRunner, BigEndianAscendingWordSerializerTest, TestCSVLoader, TestExecutePlanAction, StatsComponentTest, AtomicUpdateProcessorFactoryTest, RegexBytesRefFilterTest, HdfsUnloadDistributedZkTest, HealthCheckHandlerTest, SolrCoreMetricManagerTest, TestRequestStatusCollectionAPI, BadCopyFieldTest, SolrXmlInZkTest, TestRemoteStreaming, IndexBasedSpellCheckerTest, TestStressCloudBlindAtomicUpdates, TestPrepRecovery, TestUtilizeNode, RequestHandlersTest, TestObjectReleaseTracker, TestDistribIDF, HdfsAutoAddReplicasIntegrationTest, SchemaApiFailureTest, FieldAnalysisRequestHandlerTest, SystemInfoHandlerTest, DistributedQueryElevationComponentTest, BJQParserTest, MBeansHandlerTest, CollectionsAPIAsyncDistributedZkTest, DocumentAnalysisRequestHandlerTest, AutoScalingHandlerTest, PKIAuthenticationIntegrationTest, TestRawTransformer, XmlInterpolationTest, TestSolrDeletionPolicy1, TestExportWriter, TestManagedStopFilterFactory, TestQuerySenderNoQuery, MetricsHistoryHandlerTest, TestDocBasedVersionConstraints, NodeMarkersRegistrationTest, FullSolrCloudDistribCmdsTest, TestCloudDeleteByQuery, TestSolrCLIRunExample, ShufflingReplicaListTransformerTest, AssignTest, BlobRepositoryCloudTest, TestCrossCoreJoin, HDFSCollectionsAPITest, HttpSolrCallGetCoreTest, TestLeaderElectionWithEmptyReplica, TestLuceneMatchVersion, TestLegacyNumericUtils, HdfsChaosMonkeyNothingIsSafeTest, ConfigureRecoveryStrategyTest, TestIBSimilarityFactory, TestEmbeddedSolrServerAdminHandler, HdfsSyncSliceTest, CoreMergeIndexesAdminHandlerTest, CustomCollectionTest, TestConfig, SimpleFacetsTest, OverseerStatusTest, HighlighterConfigTest, RollingRestartTest, TestSolrQueryParser, TestDocTermOrdsUninvertLimit, TestDistributedMap, SolrCloudReportersTest, TestXmlQParser, TestIndexSearcher, TestGroupingSearch, TestCustomStream, SharedFSAutoReplicaFailoverTest, RAMDirectoryFactoryTest, TestJsonRequest, ProtectedTermFilterFactoryTest, TestJsonFacetRefinement, CdcrBidirectionalTest]
[junit4] Completed [310/821 (1!)] on J0 in 34.51s, 1 test, 1 error <<< FAILURES!
[...truncated 45707 lines...]
[repro] Jenkins log URL: https://jenkins.thetaphi.de/job/Lucene-Solr-master-Linux/22301/consoleText
[repro] Revision: 3b9d3a760a432b97aad2c08b2f778fa2344eb14a
[repro] Ant options: "-Dargs=-XX:-UseCompressedOops -XX:+UseSerialGC"
[repro] Repro line: ant test -Dtestcase=CdcrBidirectionalTest -Dtests.method=testBiDir -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[repro] ant clean
[...truncated 6 lines...]
[repro] Test suites by module:
[repro] solr/core
[repro] CdcrBidirectionalTest
[repro] ant compile-test
[...truncated 3305 lines...]
[repro] ant test-nocompile -Dtests.dups=5 -Dtests.maxfailures=5 -Dtests.class="*.CdcrBidirectionalTest" -Dtests.showOutput=onerror "-Dargs=-XX:-UseCompressedOops -XX:+UseSerialGC" -Dtests.seed=27F8202107443866 -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=bn-IN -Dtests.timezone=Asia/Tomsk -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[...truncated 76 lines...]
[repro] Failures:
[repro] 0/5 failed: org.apache.solr.cloud.cdcr.CdcrBidirectionalTest
[repro] Exiting with code 0
[...truncated 40 lines...]