You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2018/01/02 03:51:20 UTC

[JENKINS] Lucene-Solr-7.x-MacOSX (64bit/jdk-9) - Build # 374 - Unstable!

Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-MacOSX/374/
Java: 64bit/jdk-9 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  org.apache.solr.cloud.DocValuesNotIndexedTest.testGroupingDVOnly

Error Message:
Unexpected number of elements in the group for intGSF: 8

Stack Trace:
java.lang.AssertionError: Unexpected number of elements in the group for intGSF: 8
	at __randomizedtesting.SeedInfo.seed([5519E58137B423E9:CEA28BD97AEC11B7]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.apache.solr.cloud.DocValuesNotIndexedTest.testGroupingDVOnly(DocValuesNotIndexedTest.java:377)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.base/java.lang.reflect.Method.invoke(Method.java:564)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 1737 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/core/test/temp/junit4-J0-20180102_022101_79811822868173706394422.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/core/test/temp/junit4-J1-20180102_022101_79816996037534592868875.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 288 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/test-framework/test/temp/junit4-J0-20180102_022903_38113447849058799508228.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/test-framework/test/temp/junit4-J1-20180102_022903_3815801833311754988693.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 1057 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/common/test/temp/junit4-J0-20180102_022954_9669597570907608822471.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/common/test/temp/junit4-J1-20180102_022954_9663696354717594133620.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 229 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/icu/test/temp/junit4-J1-20180102_023235_1875844235838912762526.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/icu/test/temp/junit4-J0-20180102_023235_18714345580411310270729.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 243 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/kuromoji/test/temp/junit4-J0-20180102_023248_2475236165691824343067.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 9 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/kuromoji/test/temp/junit4-J1-20180102_023248_24716606010152625687614.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 154 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/morfologik/test/temp/junit4-J0-20180102_023318_3629210860932468140683.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/morfologik/test/temp/junit4-J1-20180102_023318_36210789779412425024898.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 161 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/opennlp/test/temp/junit4-J1-20180102_023324_6683829103192695052156.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/opennlp/test/temp/junit4-J0-20180102_023324_6637358382247091808771.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 168 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/phonetic/test/temp/junit4-J1-20180102_023333_2633740767143979387330.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/phonetic/test/temp/junit4-J0-20180102_023333_263566046008866856757.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 151 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/smartcn/test/temp/junit4-J0-20180102_023344_39613025712620949427448.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/smartcn/test/temp/junit4-J1-20180102_023344_3969999522883618721313.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 156 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/stempel/test/temp/junit4-J1-20180102_023352_2625002387940112592024.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/stempel/test/temp/junit4-J0-20180102_023352_2628753460539809007326.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 167 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/uima/test/temp/junit4-J0-20180102_023356_2733356000690815145193.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 191 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/backward-codecs/test/temp/junit4-J1-20180102_023412_76612657942829624837635.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 31 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/backward-codecs/test/temp/junit4-J0-20180102_023412_7669268896285142492242.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1317 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/benchmark/test/temp/junit4-J0-20180102_023556_08413615095865994666361.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/benchmark/test/temp/junit4-J1-20180102_023556_0849762698725290848535.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 234 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/classification/test/temp/junit4-J1-20180102_023606_0465164254948013469877.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/classification/test/temp/junit4-J0-20180102_023606_0461552181587111713804.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 262 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/codecs/test/temp/junit4-J1-20180102_023621_9337352792563831087294.syserr
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/codecs/test/temp/junit4-J0-20180102_023621_9335852448321232272220.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 215 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/demo/test/temp/junit4-J0-20180102_023745_15618240410015940306339.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/demo/test/temp/junit4-J1-20180102_023745_1565051482527147120731.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 170 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/expressions/test/temp/junit4-J0-20180102_023748_9181542871237744611750.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/expressions/test/temp/junit4-J1-20180102_023748_9185300388899852634364.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 228 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/facet/test/temp/junit4-J1-20180102_023754_8085027016578967383039.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/facet/test/temp/junit4-J0-20180102_023754_8086282543811191810280.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 179 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/grouping/test/temp/junit4-J1-20180102_023845_1627478037109934858610.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/grouping/test/temp/junit4-J0-20180102_023845_1623843855069000439151.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 243 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/highlighter/test/temp/junit4-J1-20180102_023853_96114353231112481519268.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/highlighter/test/temp/junit4-J0-20180102_023853_9615057395601324315823.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 163 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/join/test/temp/junit4-J1-20180102_023909_48417446645005373036493.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/join/test/temp/junit4-J0-20180102_023909_48413579307838491028460.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 147 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/memory/test/temp/junit4-J0-20180102_023921_84914210742401166343978.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/memory/test/temp/junit4-J1-20180102_023921_8493221154086140856993.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 184 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/misc/test/temp/junit4-J1-20180102_023927_9635180682508986343244.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/misc/test/temp/junit4-J0-20180102_023927_96314590830732868873328.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 329 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/queries/test/temp/junit4-J1-20180102_023948_57618360287428337508223.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/queries/test/temp/junit4-J0-20180102_023948_5761160763889292415098.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 227 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/queryparser/test/temp/junit4-J1-20180102_024000_5882068974359308721779.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/queryparser/test/temp/junit4-J0-20180102_024000_58615447483225793157267.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 187 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/replicator/test/temp/junit4-J1-20180102_024009_09515173296340421243327.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 23 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/replicator/test/temp/junit4-J0-20180102_024009_09517260801063412995229.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 202 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/sandbox/test/temp/junit4-J1-20180102_024022_3844373369387277358719.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 21 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/sandbox/test/temp/junit4-J0-20180102_024022_38416837704333775052307.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 317 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial-extras/test/temp/junit4-J0-20180102_024101_38316484233977681560386.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial-extras/test/temp/junit4-J1-20180102_024101_3838063604472603393023.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 146 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial/test/temp/junit4-J0-20180102_024120_48818241826212917870131.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 170 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial3d/test/temp/junit4-J0-20180102_024123_08515150752197652867379.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 14 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial3d/test/temp/junit4-J1-20180102_024123_08514672341126993353969.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 250 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/suggest/test/temp/junit4-J0-20180102_024401_5921779957371626415542.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/suggest/test/temp/junit4-J1-20180102_024401_5925239396156889366558.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3367 lines...]
   [junit4] Suite: org.apache.solr.cloud.DocValuesNotIndexedTest
   [junit4]   2> 2266641 INFO  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> Creating dataDir: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/init-core-data-001
   [junit4]   2> 2266641 WARN  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=7 numCloses=7
   [junit4]   2> 2266642 INFO  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 2266642 INFO  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0) w/ MAC_OS_X supressed clientAuth
   [junit4]   2> 2266643 INFO  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001
   [junit4]   2> 2266643 INFO  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 2266643 INFO  (Thread-5193) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 2266644 INFO  (Thread-5193) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 2266646 ERROR (Thread-5193) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 2266754 INFO  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.c.ZkTestServer start zk server on port:57153
   [junit4]   2> 2266807 INFO  (zkConnectionManagerCallback-5695-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2266811 INFO  (jetty-launcher-5692-thread-2) [    ] o.e.j.s.Server jetty-9.3.20.v20170531
   [junit4]   2> 2266811 INFO  (jetty-launcher-5692-thread-1) [    ] o.e.j.s.Server jetty-9.3.20.v20170531
   [junit4]   2> 2266814 INFO  (jetty-launcher-5692-thread-2) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@4ba33d92{/solr,null,AVAILABLE}
   [junit4]   2> 2266814 INFO  (jetty-launcher-5692-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@15a92eb8{/solr,null,AVAILABLE}
   [junit4]   2> 2266815 INFO  (jetty-launcher-5692-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@66798e88{HTTP/1.1,[http/1.1]}{127.0.0.1:57156}
   [junit4]   2> 2266816 INFO  (jetty-launcher-5692-thread-1) [    ] o.e.j.s.Server Started @2276309ms
   [junit4]   2> 2266816 INFO  (jetty-launcher-5692-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=57156}
   [junit4]   2> 2266816 INFO  (jetty-launcher-5692-thread-2) [    ] o.e.j.s.AbstractConnector Started ServerConnector@bc50b82{HTTP/1.1,[http/1.1]}{127.0.0.1:57155}
   [junit4]   2> 2266816 INFO  (jetty-launcher-5692-thread-2) [    ] o.e.j.s.Server Started @2276310ms
   [junit4]   2> 2266816 ERROR (jetty-launcher-5692-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 2266816 INFO  (jetty-launcher-5692-thread-2) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=57155}
   [junit4]   2> 2266816 INFO  (jetty-launcher-5692-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 7.3.0
   [junit4]   2> 2266816 ERROR (jetty-launcher-5692-thread-2) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 2266816 INFO  (jetty-launcher-5692-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 2266817 INFO  (jetty-launcher-5692-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 2266817 INFO  (jetty-launcher-5692-thread-2) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 7.3.0
   [junit4]   2> 2266817 INFO  (jetty-launcher-5692-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-01-02T03:22:57.478226Z
   [junit4]   2> 2266817 INFO  (jetty-launcher-5692-thread-2) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 2266817 INFO  (jetty-launcher-5692-thread-2) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 2266817 INFO  (jetty-launcher-5692-thread-2) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-01-02T03:22:57.478391Z
   [junit4]   2> 2266818 INFO  (zkConnectionManagerCallback-5697-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2266818 INFO  (zkConnectionManagerCallback-5699-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2266819 INFO  (jetty-launcher-5692-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 2266819 INFO  (jetty-launcher-5692-thread-2) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 2266824 INFO  (jetty-launcher-5692-thread-2) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:57153/solr
   [junit4]   2> 2266825 INFO  (jetty-launcher-5692-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:57153/solr
   [junit4]   2> 2266826 INFO  (zkConnectionManagerCallback-5705-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2266826 INFO  (zkConnectionManagerCallback-5707-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2266828 INFO  (zkConnectionManagerCallback-5709-thread-1-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2266829 INFO  (zkConnectionManagerCallback-5711-thread-1-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2266980 INFO  (jetty-launcher-5692-thread-2) [n:127.0.0.1:57155_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 2266980 INFO  (jetty-launcher-5692-thread-1) [n:127.0.0.1:57156_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 2266981 INFO  (jetty-launcher-5692-thread-2) [n:127.0.0.1:57155_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:57155_solr
   [junit4]   2> 2266982 INFO  (jetty-launcher-5692-thread-1) [n:127.0.0.1:57156_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:57156_solr
   [junit4]   2> 2266983 INFO  (jetty-launcher-5692-thread-2) [n:127.0.0.1:57155_solr    ] o.a.s.c.Overseer Overseer (id=99278086295191557-127.0.0.1:57155_solr-n_0000000000) starting
   [junit4]   2> 2266984 INFO  (zkCallback-5710-thread-1-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 2266984 INFO  (zkCallback-5708-thread-1-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 2266995 INFO  (jetty-launcher-5692-thread-2) [n:127.0.0.1:57155_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:57155_solr
   [junit4]   2> 2266997 INFO  (zkCallback-5710-thread-1-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 2266997 INFO  (zkCallback-5708-thread-1-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 2267033 INFO  (jetty-launcher-5692-thread-2) [n:127.0.0.1:57155_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57155.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2267033 INFO  (jetty-launcher-5692-thread-1) [n:127.0.0.1:57156_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57156.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2267040 INFO  (jetty-launcher-5692-thread-2) [n:127.0.0.1:57155_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57155.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2267041 INFO  (jetty-launcher-5692-thread-2) [n:127.0.0.1:57155_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57155.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2267041 INFO  (jetty-launcher-5692-thread-1) [n:127.0.0.1:57156_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57156.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2267041 INFO  (jetty-launcher-5692-thread-1) [n:127.0.0.1:57156_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57156.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2267042 INFO  (jetty-launcher-5692-thread-2) [n:127.0.0.1:57155_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node2/.
   [junit4]   2> 2267042 INFO  (jetty-launcher-5692-thread-1) [n:127.0.0.1:57156_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node1/.
   [junit4]   2> 2267074 INFO  (zkConnectionManagerCallback-5717-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2267077 INFO  (zkConnectionManagerCallback-5721-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2267078 INFO  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 2267080 INFO  (SUITE-DocValuesNotIndexedTest-seed#[5519E58137B423E9]-worker) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:57153/solr ready
   [junit4]   2> 2267086 INFO  (qtp1609785371-24981) [n:127.0.0.1:57155_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params replicationFactor=1&collection.configName=conf1&maxShardsPerNode=2&name=dv_coll&nrtReplicas=1&action=CREATE&numShards=4&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 2267089 INFO  (OverseerThreadFactory-11898-thread-1-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.CreateCollectionCmd Create collection dv_coll
   [junit4]   2> 2267214 INFO  (OverseerStateUpdate-99278086295191557-127.0.0.1:57155_solr-n_0000000000) [n:127.0.0.1:57155_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"dv_coll",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"dv_coll_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:57156/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 2267222 INFO  (OverseerStateUpdate-99278086295191557-127.0.0.1:57155_solr-n_0000000000) [n:127.0.0.1:57155_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"dv_coll",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"dv_coll_shard2_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:57155/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 2267229 INFO  (OverseerStateUpdate-99278086295191557-127.0.0.1:57155_solr-n_0000000000) [n:127.0.0.1:57155_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"dv_coll",
   [junit4]   2>   "shard":"shard3",
   [junit4]   2>   "core":"dv_coll_shard3_replica_n4",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:57156/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 2267237 INFO  (OverseerStateUpdate-99278086295191557-127.0.0.1:57155_solr-n_0000000000) [n:127.0.0.1:57155_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"dv_coll",
   [junit4]   2>   "shard":"shard4",
   [junit4]   2>   "core":"dv_coll_shard4_replica_n6",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:57155/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 2267460 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr    ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=dv_coll_shard1_replica_n1&action=CREATE&numShards=4&collection=dv_coll&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 2267460 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr    ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=dv_coll_shard3_replica_n4&action=CREATE&numShards=4&collection=dv_coll&shard=shard3&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 2267460 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 2267461 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr    ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=dv_coll_shard2_replica_n2&action=CREATE&numShards=4&collection=dv_coll&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 2267462 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr    ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=dv_coll_shard4_replica_n6&action=CREATE&numShards=4&collection=dv_coll&shard=shard4&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 2267462 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 2267582 INFO  (zkCallback-5708-thread-1-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2267583 INFO  (zkCallback-5710-thread-1-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2267583 INFO  (zkCallback-5708-thread-2-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2267584 INFO  (zkCallback-5710-thread-2-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2268511 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 2268511 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 2268511 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 2268521 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 2268526 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.IndexSchema [dv_coll_shard3_replica_n4] Schema name=minimal
   [junit4]   2> 2268528 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.IndexSchema [dv_coll_shard2_replica_n2] Schema name=minimal
   [junit4]   2> 2268528 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2268528 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard3_replica_n4' using configuration from collection dv_coll, trusted=true
   [junit4]   2> 2268529 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57156.solr.core.dv_coll.shard3.replica_n4' (registry 'solr.core.dv_coll.shard3.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2268529 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 2268529 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [[dv_coll_shard3_replica_n4] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node1/dv_coll_shard3_replica_n4], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node1/./dv_coll_shard3_replica_n4/data/]
   [junit4]   2> 2268529 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.IndexSchema [dv_coll_shard4_replica_n6] Schema name=minimal
   [junit4]   2> 2268530 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2268530 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard2_replica_n2' using configuration from collection dv_coll, trusted=true
   [junit4]   2> 2268530 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57155.solr.core.dv_coll.shard2.replica_n2' (registry 'solr.core.dv_coll.shard2.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2268530 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 2268531 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [[dv_coll_shard2_replica_n2] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node2/dv_coll_shard2_replica_n2], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node2/./dv_coll_shard2_replica_n2/data/]
   [junit4]   2> 2268531 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2268531 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard4_replica_n6' using configuration from collection dv_coll, trusted=true
   [junit4]   2> 2268531 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57155.solr.core.dv_coll.shard4.replica_n6' (registry 'solr.core.dv_coll.shard4.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2268532 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 2268532 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [[dv_coll_shard4_replica_n6] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node2/dv_coll_shard4_replica_n6], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node2/./dv_coll_shard4_replica_n6/data/]
   [junit4]   2> 2268533 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 2268535 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2268535 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard1_replica_n1' using configuration from collection dv_coll, trusted=true
   [junit4]   2> 2268535 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57156.solr.core.dv_coll.shard1.replica_n1' (registry 'solr.core.dv_coll.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2268535 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 2268535 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [[dv_coll_shard1_replica_n1] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node1/dv_coll_shard1_replica_n1], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node1/./dv_coll_shard1_replica_n1/data/]
   [junit4]   2> 2268587 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 2268587 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 2268589 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 2268589 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 2268592 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 2268592 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 2268594 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 2268594 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 2268595 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@496b0b97[dv_coll_shard3_replica_n4] main]
   [junit4]   2> 2268596 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 2268597 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 2268597 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 2268597 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@1e91c7be[dv_coll_shard2_replica_n2] main]
   [junit4]   2> 2268597 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 2268598 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 2268598 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588449382769885184
   [junit4]   2> 2268599 INFO  (searcherExecutor-11903-thread-1-processing-n:127.0.0.1:57156_solr x:dv_coll_shard3_replica_n4 s:shard3 c:dv_coll r:core_node7) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4] Registered new searcher Searcher@496b0b97[dv_coll_shard3_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 2268599 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 2268599 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 2268600 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 2268600 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 2268600 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 2268601 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 2268601 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 2268601 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 2268603 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 2268603 INFO  (searcherExecutor-11904-thread-1-processing-n:127.0.0.1:57155_solr x:dv_coll_shard2_replica_n2 s:shard2 c:dv_coll r:core_node5) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [dv_coll_shard2_replica_n2] Registered new searcher Searcher@1e91c7be[dv_coll_shard2_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 2268603 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 2268603 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 2268603 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588449382775128064
   [junit4]   2> 2268603 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:57156/solr/dv_coll_shard3_replica_n4/
   [junit4]   2> 2268603 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 2268604 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy http://127.0.0.1:57156/solr/dv_coll_shard3_replica_n4/ has no replicas
   [junit4]   2> 2268604 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 2268607 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@68994ed5[dv_coll_shard4_replica_n6] main]
   [junit4]   2> 2268608 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:57156/solr/dv_coll_shard3_replica_n4/ shard3
   [junit4]   2> 2268608 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 2268609 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 2268609 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 2268610 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 2268610 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 2268610 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:57155/solr/dv_coll_shard2_replica_n2/
   [junit4]   2> 2268610 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 2268610 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy http://127.0.0.1:57155/solr/dv_coll_shard2_replica_n2/ has no replicas
   [junit4]   2> 2268610 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 2268610 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 2268611 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 2268611 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588449382783516672
   [junit4]   2> 2268612 INFO  (searcherExecutor-11905-thread-1-processing-n:127.0.0.1:57155_solr x:dv_coll_shard4_replica_n6 s:shard4 c:dv_coll r:core_node8) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [dv_coll_shard4_replica_n6] Registered new searcher Searcher@68994ed5[dv_coll_shard4_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 2268615 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:57155/solr/dv_coll_shard2_replica_n2/ shard2
   [junit4]   2> 2268616 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 2268616 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 2268616 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:57155/solr/dv_coll_shard4_replica_n6/
   [junit4]   2> 2268616 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 2268616 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy http://127.0.0.1:57155/solr/dv_coll_shard4_replica_n6/ has no replicas
   [junit4]   2> 2268616 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 2268620 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:57155/solr/dv_coll_shard4_replica_n6/ shard4
   [junit4]   2> 2268624 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 2268624 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 2268625 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 2268625 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 2268626 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@1ae6f2b6[dv_coll_shard1_replica_n1] main]
   [junit4]   2> 2268626 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 2268630 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 2268630 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 2268631 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 2268632 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 2268633 INFO  (searcherExecutor-11906-thread-1-processing-n:127.0.0.1:57156_solr x:dv_coll_shard1_replica_n1 s:shard1 c:dv_coll r:core_node3) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@1ae6f2b6[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 2268634 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588449382807633920
   [junit4]   2> 2268639 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 2268639 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 2268639 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:57156/solr/dv_coll_shard1_replica_n1/
   [junit4]   2> 2268639 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 2268639 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy http://127.0.0.1:57156/solr/dv_coll_shard1_replica_n1/ has no replicas
   [junit4]   2> 2268640 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 2268644 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:57156/solr/dv_coll_shard1_replica_n1/ shard1
   [junit4]   2> 2268756 INFO  (zkCallback-5708-thread-2-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2268757 INFO  (zkCallback-5710-thread-1-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2268756 INFO  (zkCallback-5710-thread-2-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2268756 INFO  (zkCallback-5708-thread-3-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2268781 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 2268789 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 2268801 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 2268804 INFO  (qtp2032969802-24980) [n:127.0.0.1:57156_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=dv_coll_shard3_replica_n4&action=CREATE&numShards=4&collection=dv_coll&shard=shard3&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1344
   [junit4]   2> 2268807 INFO  (qtp1609785371-24983) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=dv_coll_shard2_replica_n2&action=CREATE&numShards=4&collection=dv_coll&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1345
   [junit4]   2> 2268808 INFO  (qtp1609785371-24984) [n:127.0.0.1:57155_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=dv_coll_shard4_replica_n6&action=CREATE&numShards=4&collection=dv_coll&shard=shard4&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1346
   [junit4]   2> 2268816 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 2268820 INFO  (qtp2032969802-24978) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=dv_coll_shard1_replica_n1&action=CREATE&numShards=4&collection=dv_coll&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1360
   [junit4]   2> 2268826 INFO  (qtp1609785371-24981) [n:127.0.0.1:57155_solr    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 2268925 INFO  (zkCallback-5710-thread-2-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2268925 INFO  (zkCallback-5710-thread-1-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2268928 INFO  (zkCallback-5708-thread-3-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2268928 INFO  (zkCallback-5708-thread-2-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 2269092 INFO  (OverseerCollectionConfigSetProcessor-99278086295191557-127.0.0.1:57155_solr-n_0000000000) [n:127.0.0.1:57155_solr    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 2269836 INFO  (qtp1609785371-24981) [n:127.0.0.1:57155_solr    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={replicationFactor=1&collection.configName=conf1&maxShardsPerNode=2&name=dv_coll&nrtReplicas=1&action=CREATE&numShards=4&wt=javabin&version=2} status=0 QTime=2749
   [junit4]   2> 2269846 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SchemaManager managed schema loaded . version : 0 
   [junit4]   2> 2269848 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 2269851 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2269860 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ZkController Persisted config data to node /configs/conf1/managed-schema 
   [junit4]   2> 2269860 INFO  (zkCallback-5710-thread-1-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 2269860 INFO  (zkCallback-5710-thread-2-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 2269860 INFO  (zkCallback-5708-thread-2-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 2269861 INFO  (zkCallback-5708-thread-3-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 2269863 INFO  (zkCallback-5708-thread-2-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
   [junit4]   2> 2269863 INFO  (zkCallback-5710-thread-2-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
   [junit4]   2> 2269865 INFO  (zkCallback-5708-thread-2-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.IndexSchema [dv_coll_shard4_replica_n6] Schema name=minimal
   [junit4]   2> 2269865 INFO  (zkCallback-5710-thread-2-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.s.IndexSchema [dv_coll_shard3_replica_n4] Schema name=minimal
   [junit4]   2> 2269868 INFO  (zkCallback-5708-thread-3-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
   [junit4]   2> 2269888 INFO  (zkCallback-5708-thread-2-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2269888 INFO  (zkCallback-5708-thread-2-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 25 ms
   [junit4]   2> 2269888 INFO  (zkCallback-5708-thread-3-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.IndexSchema [dv_coll_shard2_replica_n2] Schema name=minimal
   [junit4]   2> 2269889 INFO  (zkCallback-5710-thread-2-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2269889 INFO  (zkCallback-5710-thread-2-processing-n:127.0.0.1:57156_solr) [n:127.0.0.1:57156_solr    ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 25 ms
   [junit4]   2> 2269890 INFO  (Thread-5212) [n:127.0.0.1:57155_solr    ] o.a.s.c.SolrCore config update listener called for core dv_coll_shard2_replica_n2
   [junit4]   2> 2269891 INFO  (Thread-5213) [n:127.0.0.1:57156_solr    ] o.a.s.c.SolrCore config update listener called for core dv_coll_shard1_replica_n1
   [junit4]   2> 2269891 INFO  (zkCallback-5708-thread-3-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2269891 INFO  (zkCallback-5708-thread-3-processing-n:127.0.0.1:57155_solr) [n:127.0.0.1:57155_solr    ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 4 ms
   [junit4]   2> 2269893 INFO  (Thread-5212) [n:127.0.0.1:57155_solr    ] o.a.s.c.SolrCore core reload dv_coll_shard2_replica_n2
   [junit4]   2> 2269894 INFO  (Thread-5213) [n:127.0.0.1:57156_solr    ] o.a.s.c.SolrCore core reload dv_coll_shard1_replica_n1
   [junit4]   2> 2269898 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 2269901 INFO  (Thread-5212) [n:127.0.0.1:57155_solr    ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 2269904 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 2269924 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2269924 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.CoreContainer Reloading SolrCore 'dv_coll_shard1_replica_n1' using configuration from collection dv_coll
   [junit4]   2> 2269948 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57156.solr.core.dv_coll.shard1.replica_n1' (registry 'solr.core.dv_coll.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2269948 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [[dv_coll_shard1_replica_n1] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node1/dv_coll_shard1_replica_n1], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node1/./dv_coll_shard1_replica_n1/data/]
   [junit4]   2> 2269952 INFO  (Thread-5212) [n:127.0.0.1:57155_solr    ] o.a.s.s.IndexSchema [dv_coll_shard2_replica_n2] Schema name=minimal
   [junit4]   2> 2269954 INFO  (Thread-5212) [n:127.0.0.1:57155_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 2269954 INFO  (Thread-5212) [n:127.0.0.1:57155_solr    ] o.a.s.c.CoreContainer Reloading SolrCore 'dv_coll_shard2_replica_n2' using configuration from collection dv_coll
   [junit4]   2> 2269973 INFO  (Thread-5212) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_57155.solr.core.dv_coll.shard2.replica_n2' (registry 'solr.core.dv_coll.shard2.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@32c1e1e4
   [junit4]   2> 2269973 INFO  (Thread-5212) [n:127.0.0.1:57155_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [[dv_coll_shard2_replica_n2] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node2/dv_coll_shard2_replica_n2], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J0/temp/solr.cloud.DocValuesNotIndexedTest_5519E58137B423E9-001/tempDir-001/node2/./dv_coll_shard2_replica_n2/data/]
   [junit4]   2> 2270003 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 2270003 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 2270006 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@32a5d662[dv_coll_shard1_replica_n1] main]
   [junit4]   2> 2270008 INFO  (qtp2032969802-24986) [n:127.0.0.1:57156_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStora

[...truncated too long message...]

VM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-solrj/test/temp/junit4-J0-20180102_033843_22117816717016854801957.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-solrj/test/temp/junit4-J1-20180102_033843_22116464354923148765914.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 1463 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-analysis-extras/test/temp/junit4-J1-20180102_034226_83015020158076919136211.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-analysis-extras/test/temp/junit4-J0-20180102_034226_8306369309173161031768.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 782 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-analytics/test/temp/junit4-J0-20180102_034242_68418371683708102109826.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-analytics/test/temp/junit4-J1-20180102_034242_6848946336296615589203.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 553 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-clustering/test/temp/junit4-J1-20180102_034349_0486982427934938825406.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-clustering/test/temp/junit4-J0-20180102_034349_048576185824784518701.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1258 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-dataimporthandler-extras/test/temp/junit4-J1-20180102_034414_9583614708304055958837.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-dataimporthandler-extras/test/temp/junit4-J0-20180102_034414_95818055145524216562140.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 646 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-dataimporthandler/test/temp/junit4-J0-20180102_034428_87814364721121284204309.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-dataimporthandler/test/temp/junit4-J1-20180102_034428_87812770919458027069094.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 558 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-cell/test/temp/junit4-J0-20180102_034501_3821975109377818347918.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-cell/test/temp/junit4-J1-20180102_034501_3827972494755314770383.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 561 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-langid/test/temp/junit4-J1-20180102_034521_38414219771780520198177.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-langid/test/temp/junit4-J0-20180102_034521_3848095926278349213403.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 654 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-ltr/test/temp/junit4-J0-20180102_034533_2151971658431540288905.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-ltr/test/temp/junit4-J1-20180102_034533_21516832887161273791324.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 595 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-uima/test/temp/junit4-J1-20180102_034610_03018255168935704078831.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-uima/test/temp/junit4-J0-20180102_034610_03013464751816726544796.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 548 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-velocity/test/temp/junit4-J0-20180102_034625_5477438421391335379007.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 32273 lines...]

[JENKINS] Lucene-Solr-7.x-MacOSX (64bit/jdk-9) - Build # 376 - Still Unstable!

Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-MacOSX/376/
Java: 64bit/jdk-9 -XX:-UseCompressedOops -XX:+UseParallelGC

1 tests failed.
FAILED:  org.apache.solr.cloud.DocValuesNotIndexedTest.testGroupingDVOnly

Error Message:
Unexpected number of elements in the group for intGSF: 8

Stack Trace:
java.lang.AssertionError: Unexpected number of elements in the group for intGSF: 8
	at __randomizedtesting.SeedInfo.seed([CDE0861F80EE206D:565BE847CDB61233]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.apache.solr.cloud.DocValuesNotIndexedTest.testGroupingDVOnly(DocValuesNotIndexedTest.java:377)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.base/java.lang.reflect.Method.invoke(Method.java:564)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 11633 lines...]
   [junit4] Suite: org.apache.solr.cloud.DocValuesNotIndexedTest
   [junit4]   2> Creating dataDir: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/init-core-data-001
   [junit4]   2> 0    INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 87   INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.e.j.u.log Logging initialized @1855ms
   [junit4]   2> 96   INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0) w/ MAC_OS_X supressed clientAuth
   [junit4]   2> 103  INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 133  INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 2 servers in /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001
   [junit4]   2> 146  INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 148  INFO  (Thread-0) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 148  INFO  (Thread-0) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 252  INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.c.ZkTestServer start zk server on port:65145
   [junit4]   2> 335  WARN  (NIOServerCxn.Factory:0.0.0.0/0.0.0.0:0) [    ] o.a.z.s.NIOServerCnxn Exception causing close of session 0x0 due to java.io.IOException: ZooKeeperServer not running
   [junit4]   2> 336  ERROR (Thread-0) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 1935 INFO  (zkConnectionManagerCallback-4-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2111 INFO  (jetty-launcher-1-thread-2) [    ] o.e.j.s.Server jetty-9.3.20.v20170531
   [junit4]   2> 2112 INFO  (jetty-launcher-1-thread-1) [    ] o.e.j.s.Server jetty-9.3.20.v20170531
   [junit4]   2> 2130 INFO  (jetty-launcher-1-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@2694c442{/solr,null,AVAILABLE}
   [junit4]   2> 2130 INFO  (jetty-launcher-1-thread-2) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@7ae8e815{/solr,null,AVAILABLE}
   [junit4]   2> 2135 INFO  (jetty-launcher-1-thread-2) [    ] o.e.j.s.AbstractConnector Started ServerConnector@294bd6d8{HTTP/1.1,[http/1.1]}{127.0.0.1:65157}
   [junit4]   2> 2135 INFO  (jetty-launcher-1-thread-2) [    ] o.e.j.s.Server Started @3911ms
   [junit4]   2> 2135 INFO  (jetty-launcher-1-thread-2) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=65157}
   [junit4]   2> 2135 INFO  (jetty-launcher-1-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@2d799b06{HTTP/1.1,[http/1.1]}{127.0.0.1:65156}
   [junit4]   2> 2136 INFO  (jetty-launcher-1-thread-1) [    ] o.e.j.s.Server Started @3912ms
   [junit4]   2> 2136 INFO  (jetty-launcher-1-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=65156}
   [junit4]   2> 2188 ERROR (jetty-launcher-1-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 2189 ERROR (jetty-launcher-1-thread-2) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 2189 INFO  (jetty-launcher-1-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 7.3.0
   [junit4]   2> 2190 INFO  (jetty-launcher-1-thread-2) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr™ version 7.3.0
   [junit4]   2> 2225 INFO  (jetty-launcher-1-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 2225 INFO  (jetty-launcher-1-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 2226 INFO  (jetty-launcher-1-thread-2) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 2226 INFO  (jetty-launcher-1-thread-2) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 2262 INFO  (jetty-launcher-1-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-01-03T00:37:44.086270Z
   [junit4]   2> 2263 INFO  (jetty-launcher-1-thread-2) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-01-03T00:37:44.086303Z
   [junit4]   2> 2333 INFO  (jetty-launcher-1-thread-2) [    ] o.a.s.c.SolrResourceLoader solr home defaulted to 'solr/' (could not find system property or JNDI)
   [junit4]   2> 2336 INFO  (zkConnectionManagerCallback-6-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2338 INFO  (zkConnectionManagerCallback-8-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2339 INFO  (jetty-launcher-1-thread-2) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 2340 INFO  (jetty-launcher-1-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 2860 INFO  (jetty-launcher-1-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:65145/solr
   [junit4]   2> 2860 INFO  (jetty-launcher-1-thread-2) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:65145/solr
   [junit4]   2> 2889 INFO  (zkConnectionManagerCallback-16-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2890 INFO  (zkConnectionManagerCallback-14-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2916 INFO  (zkConnectionManagerCallback-18-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 2926 INFO  (zkConnectionManagerCallback-20-thread-1-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 3212 INFO  (jetty-launcher-1-thread-1) [n:127.0.0.1:65156_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 3212 INFO  (jetty-launcher-1-thread-2) [n:127.0.0.1:65157_solr    ] o.a.s.c.Overseer Overseer (id=null) closing
   [junit4]   2> 3216 INFO  (jetty-launcher-1-thread-1) [n:127.0.0.1:65156_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:65156_solr
   [junit4]   2> 3217 INFO  (jetty-launcher-1-thread-2) [n:127.0.0.1:65157_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:65157_solr
   [junit4]   2> 3220 INFO  (jetty-launcher-1-thread-1) [n:127.0.0.1:65156_solr    ] o.a.s.c.Overseer Overseer (id=99283098800095238-127.0.0.1:65156_solr-n_0000000000) starting
   [junit4]   2> 3232 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 3232 INFO  (zkCallback-19-thread-1-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 3462 INFO  (jetty-launcher-1-thread-1) [n:127.0.0.1:65156_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:65156_solr
   [junit4]   2> 3469 INFO  (zkCallback-19-thread-1-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 3470 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 3896 INFO  (jetty-launcher-1-thread-1) [n:127.0.0.1:65156_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65156.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 3897 INFO  (jetty-launcher-1-thread-1) [n:127.0.0.1:65156_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65156.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 3897 INFO  (jetty-launcher-1-thread-1) [n:127.0.0.1:65156_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65156.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 3897 INFO  (jetty-launcher-1-thread-2) [n:127.0.0.1:65157_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65157.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 3902 INFO  (jetty-launcher-1-thread-2) [n:127.0.0.1:65157_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65157.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 3902 INFO  (jetty-launcher-1-thread-2) [n:127.0.0.1:65157_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65157.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 3932 INFO  (jetty-launcher-1-thread-2) [n:127.0.0.1:65157_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/.
   [junit4]   2> 3933 INFO  (jetty-launcher-1-thread-1) [n:127.0.0.1:65156_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node1/.
   [junit4]   2> 4057 INFO  (zkConnectionManagerCallback-26-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 4064 INFO  (zkConnectionManagerCallback-30-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 4067 INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 4069 INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:65145/solr ready
   [junit4]   2> 4313 INFO  (qtp655761501-28) [n:127.0.0.1:65156_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params replicationFactor=1&collection.configName=conf1&maxShardsPerNode=2&name=dv_coll&nrtReplicas=1&action=CREATE&numShards=4&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 4326 INFO  (OverseerThreadFactory-12-thread-1-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.CreateCollectionCmd Create collection dv_coll
   [junit4]   2> 4464 INFO  (OverseerStateUpdate-99283098800095238-127.0.0.1:65156_solr-n_0000000000) [n:127.0.0.1:65156_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"dv_coll",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"dv_coll_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:65157/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 4467 INFO  (OverseerStateUpdate-99283098800095238-127.0.0.1:65156_solr-n_0000000000) [n:127.0.0.1:65156_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"dv_coll",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"dv_coll_shard2_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:65156/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 4472 INFO  (OverseerStateUpdate-99283098800095238-127.0.0.1:65156_solr-n_0000000000) [n:127.0.0.1:65156_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"dv_coll",
   [junit4]   2>   "shard":"shard3",
   [junit4]   2>   "core":"dv_coll_shard3_replica_n4",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:65157/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 4476 INFO  (OverseerStateUpdate-99283098800095238-127.0.0.1:65156_solr-n_0000000000) [n:127.0.0.1:65156_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"dv_coll",
   [junit4]   2>   "shard":"shard4",
   [junit4]   2>   "core":"dv_coll_shard4_replica_n6",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:65156/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 4711 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr    ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=dv_coll_shard4_replica_n6&action=CREATE&numShards=4&collection=dv_coll&shard=shard4&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 4717 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 4715 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr    ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=dv_coll_shard2_replica_n2&action=CREATE&numShards=4&collection=dv_coll&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 4716 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr    ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=dv_coll_shard1_replica_n1&action=CREATE&numShards=4&collection=dv_coll&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 4724 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr    ] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=dv_coll_shard3_replica_n4&action=CREATE&numShards=4&collection=dv_coll&shard=shard3&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 4724 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 4877 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 4877 INFO  (zkCallback-19-thread-1-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 4878 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 4879 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 5804 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 5805 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 5810 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 5820 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 5888 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.IndexSchema [dv_coll_shard2_replica_n2] Schema name=minimal
   [junit4]   2> 5892 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.IndexSchema [dv_coll_shard4_replica_n6] Schema name=minimal
   [junit4]   2> 5896 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.IndexSchema [dv_coll_shard3_replica_n4] Schema name=minimal
   [junit4]   2> 5904 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 5927 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 5927 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard1_replica_n1' using configuration from collection dv_coll, trusted=true
   [junit4]   2> 5928 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 5928 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard4_replica_n6' using configuration from collection dv_coll, trusted=true
   [junit4]   2> 5928 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 5928 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard3_replica_n4' using configuration from collection dv_coll, trusted=true
   [junit4]   2> 5934 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 5935 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'dv_coll_shard2_replica_n2' using configuration from collection dv_coll, trusted=true
   [junit4]   2> 5942 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65157.solr.core.dv_coll.shard1.replica_n1' (registry 'solr.core.dv_coll.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 5946 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 5963 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [[dv_coll_shard1_replica_n1] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/dv_coll_shard1_replica_n1], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/./dv_coll_shard1_replica_n1/data/]
   [junit4]   2> 5965 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65157.solr.core.dv_coll.shard3.replica_n4' (registry 'solr.core.dv_coll.shard3.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 5956 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65156.solr.core.dv_coll.shard2.replica_n2' (registry 'solr.core.dv_coll.shard2.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 5967 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 5967 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [[dv_coll_shard2_replica_n2] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node1/dv_coll_shard2_replica_n2], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node1/./dv_coll_shard2_replica_n2/data/]
   [junit4]   2> 5947 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65156.solr.core.dv_coll.shard4.replica_n6' (registry 'solr.core.dv_coll.shard4.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 5966 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 5969 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 5969 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [[dv_coll_shard3_replica_n4] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/dv_coll_shard3_replica_n4], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/./dv_coll_shard3_replica_n4/data/]
   [junit4]   2> 5969 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [[dv_coll_shard4_replica_n6] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node1/dv_coll_shard4_replica_n6], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node1/./dv_coll_shard4_replica_n6/data/]
   [junit4]   2> 6525 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 6525 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 6525 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 6526 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 6526 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 6527 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 6525 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 6525 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 6545 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 6545 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 6545 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 6545 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 6545 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 6546 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 6545 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 6547 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 6567 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@1038222c[dv_coll_shard1_replica_n1] main]
   [junit4]   2> 6567 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@1eafb22a[dv_coll_shard3_replica_n4] main]
   [junit4]   2> 6567 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@61d16c1e[dv_coll_shard2_replica_n2] main]
   [junit4]   2> 6567 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@3930e5c0[dv_coll_shard4_replica_n6] main]
   [junit4]   2> 6581 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 6581 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 6583 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 6583 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 6592 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 6592 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 6593 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 6593 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 6594 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 6594 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 6594 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 6594 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 6595 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 6595 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 6595 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 6595 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 0 is already the latest
   [junit4]   2> 6607 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 6607 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 6608 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 6608 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 6676 INFO  (searcherExecutor-19-thread-1-processing-n:127.0.0.1:65156_solr x:dv_coll_shard2_replica_n2 s:shard2 c:dv_coll r:core_node5) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [dv_coll_shard2_replica_n2] Registered new searcher Searcher@61d16c1e[dv_coll_shard2_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 6676 INFO  (searcherExecutor-20-thread-1-processing-n:127.0.0.1:65157_solr x:dv_coll_shard3_replica_n4 s:shard3 c:dv_coll r:core_node7) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4] Registered new searcher Searcher@1eafb22a[dv_coll_shard3_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 6676 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588529587585810432
   [junit4]   2> 6676 INFO  (searcherExecutor-18-thread-1-processing-n:127.0.0.1:65156_solr x:dv_coll_shard4_replica_n6 s:shard4 c:dv_coll r:core_node8) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [dv_coll_shard4_replica_n6] Registered new searcher Searcher@3930e5c0[dv_coll_shard4_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 6676 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588529587585810432
   [junit4]   2> 6677 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588529587586859008
   [junit4]   2> 6684 INFO  (searcherExecutor-17-thread-1-processing-n:127.0.0.1:65157_solr x:dv_coll_shard1_replica_n1 s:shard1 c:dv_coll r:core_node3) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@1038222c[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 6687 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588529587597344768
   [junit4]   2> 6695 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 6696 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 6696 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:65157/solr/dv_coll_shard3_replica_n4/
   [junit4]   2> 6697 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 6697 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SyncStrategy http://127.0.0.1:65157/solr/dv_coll_shard3_replica_n4/ has no replicas
   [junit4]   2> 6697 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 6698 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 6705 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 6705 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:65156/solr/dv_coll_shard2_replica_n2/
   [junit4]   2> 6706 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 6707 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SyncStrategy http://127.0.0.1:65156/solr/dv_coll_shard2_replica_n2/ has no replicas
   [junit4]   2> 6707 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 6708 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:65157/solr/dv_coll_shard3_replica_n4/ shard3
   [junit4]   2> 6713 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 6717 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 6717 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:65156/solr/dv_coll_shard4_replica_n6/
   [junit4]   2> 6717 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 6718 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SyncStrategy http://127.0.0.1:65156/solr/dv_coll_shard4_replica_n6/ has no replicas
   [junit4]   2> 6718 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 6717 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:65156/solr/dv_coll_shard2_replica_n2/ shard2
   [junit4]   2> 6722 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 6722 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 6722 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:65157/solr/dv_coll_shard1_replica_n1/
   [junit4]   2> 6722 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 6722 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SyncStrategy http://127.0.0.1:65157/solr/dv_coll_shard1_replica_n1/ has no replicas
   [junit4]   2> 6722 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 6728 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:65156/solr/dv_coll_shard4_replica_n6/ shard4
   [junit4]   2> 6732 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:65157/solr/dv_coll_shard1_replica_n1/ shard1
   [junit4]   2> 6836 INFO  (zkCallback-19-thread-3-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 6836 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 6836 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 6836 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 6844 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 6848 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=dv_coll_shard1_replica_n1&action=CREATE&numShards=4&collection=dv_coll&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=2137
   [junit4]   2> 6878 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 6879 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 6886 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=dv_coll_shard3_replica_n4&action=CREATE&numShards=4&collection=dv_coll&shard=shard3&wt=javabin&version=2&replicaType=NRT} status=0 QTime=2178
   [junit4]   2> 6888 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 6890 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=dv_coll_shard2_replica_n2&action=CREATE&numShards=4&collection=dv_coll&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=2183
   [junit4]   2> 6898 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=dv_coll_shard4_replica_n6&action=CREATE&numShards=4&collection=dv_coll&shard=shard4&wt=javabin&version=2&replicaType=NRT} status=0 QTime=2194
   [junit4]   2> 6926 INFO  (qtp655761501-28) [n:127.0.0.1:65156_solr    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 7007 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 7007 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 7007 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 7007 INFO  (zkCallback-19-thread-3-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [2])
   [junit4]   2> 7931 INFO  (qtp655761501-28) [n:127.0.0.1:65156_solr    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={replicationFactor=1&collection.configName=conf1&maxShardsPerNode=2&name=dv_coll&nrtReplicas=1&action=CREATE&numShards=4&wt=javabin&version=2} status=0 QTime=3640
   [junit4]   2> 7954 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SchemaManager managed schema loaded . version : 0 
   [junit4]   2> 7958 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 7961 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 7998 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.ZkController Persisted config data to node /configs/conf1/managed-schema 
   [junit4]   2> 7998 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 7998 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 7998 INFO  (zkCallback-19-thread-3-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 7998 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.ZkIndexSchemaReader A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 7999 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
   [junit4]   2> 8001 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
   [junit4]   2> 8003 INFO  (zkCallback-19-thread-3-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
   [junit4]   2> 8006 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.IndexSchema [dv_coll_shard3_replica_n4] Schema name=minimal
   [junit4]   2> 8008 INFO  (zkCallback-19-thread-3-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.IndexSchema [dv_coll_shard2_replica_n2] Schema name=minimal
   [junit4]   2> 8025 INFO  (Thread-20) [n:127.0.0.1:65156_solr    ] o.a.s.c.SolrCore config update listener called for core dv_coll_shard4_replica_n6
   [junit4]   2> 8026 INFO  (Thread-21) [n:127.0.0.1:65157_solr    ] o.a.s.c.SolrCore config update listener called for core dv_coll_shard3_replica_n4
   [junit4]   2> 8027 INFO  (zkCallback-19-thread-3-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 8027 INFO  (zkCallback-19-thread-3-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 24 ms
   [junit4]   2> 8027 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.IndexSchema [dv_coll_shard4_replica_n6] Schema name=minimal
   [junit4]   2> 8029 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 8030 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 30 ms
   [junit4]   2> 8031 INFO  (Thread-20) [n:127.0.0.1:65156_solr    ] o.a.s.c.SolrCore core reload dv_coll_shard4_replica_n6
   [junit4]   2> 8031 INFO  (Thread-21) [n:127.0.0.1:65157_solr    ] o.a.s.c.SolrCore core reload dv_coll_shard3_replica_n4
   [junit4]   2> 8036 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 8036 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 35 ms
   [junit4]   2> 8042 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 8046 INFO  (Thread-20) [n:127.0.0.1:65156_solr    ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 8056 INFO  (Thread-21) [n:127.0.0.1:65157_solr    ] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 8078 INFO  (Thread-20) [n:127.0.0.1:65156_solr    ] o.a.s.s.IndexSchema [dv_coll_shard4_replica_n6] Schema name=minimal
   [junit4]   2> 8083 INFO  (Thread-21) [n:127.0.0.1:65157_solr    ] o.a.s.s.IndexSchema [dv_coll_shard3_replica_n4] Schema name=minimal
   [junit4]   2> 8085 INFO  (Thread-20) [n:127.0.0.1:65156_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 8085 INFO  (Thread-20) [n:127.0.0.1:65156_solr    ] o.a.s.c.CoreContainer Reloading SolrCore 'dv_coll_shard4_replica_n6' using configuration from collection dv_coll
   [junit4]   2> 8096 INFO  (Thread-21) [n:127.0.0.1:65157_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 8096 INFO  (Thread-21) [n:127.0.0.1:65157_solr    ] o.a.s.c.CoreContainer Reloading SolrCore 'dv_coll_shard3_replica_n4' using configuration from collection dv_coll
   [junit4]   2> 8122 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65156.solr.core.dv_coll.shard4.replica_n6' (registry 'solr.core.dv_coll.shard4.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 8123 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [[dv_coll_shard4_replica_n6] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node1/dv_coll_shard4_replica_n6], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node1/./dv_coll_shard4_replica_n6/data/]
   [junit4]   2> 8132 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 8138 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 8138 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.CoreContainer Reloading SolrCore 'dv_coll_shard1_replica_n1' using configuration from collection dv_coll
   [junit4]   2> 8138 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65157.solr.core.dv_coll.shard3.replica_n4' (registry 'solr.core.dv_coll.shard3.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 8139 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [[dv_coll_shard3_replica_n4] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/dv_coll_shard3_replica_n4], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/./dv_coll_shard3_replica_n4/data/]
   [junit4]   2> 8186 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 8186 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 8189 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@4686ea42[dv_coll_shard4_replica_n6] main]
   [junit4]   2> 8191 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 8192 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 8192 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 8192 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.ZkIndexSchemaReader Current schema version 1 is already the latest
   [junit4]   2> 8193 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 8200 INFO  (searcherExecutor-37-thread-1-processing-n:127.0.0.1:65156_solr x:dv_coll_shard4_replica_n6 s:shard4 c:dv_coll r:core_node8) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [dv_coll_shard4_replica_n6] Registered new searcher Searcher@4686ea42[dv_coll_shard4_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 8202 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588529589185937408
   [junit4]   2> 8205 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65157.solr.core.dv_coll.shard1.replica_n1' (registry 'solr.core.dv_coll.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 8206 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [[dv_coll_shard1_replica_n1] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/dv_coll_shard1_replica_n1], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/./dv_coll_shard1_replica_n1/data/]
   [junit4]   2> 8224 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 8224 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 8228 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@41941b18[dv_coll_shard3_replica_n4] main]
   [junit4]   2> 8229 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 8230 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 8230 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 8232 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Current schema version 1 is already the latest
   [junit4]   2> 8232 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 8233 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588529589218443264
   [junit4]   2> 8234 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DefaultSolrCoreState New IndexWriter is ready to be used.
   [junit4]   2> 8236 INFO  (searcherExecutor-38-thread-1-processing-n:127.0.0.1:65157_solr x:dv_coll_shard3_replica_n4 s:shard3 c:dv_coll r:core_node7) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4] Registered new searcher Searcher@41941b18[dv_coll_shard3_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 8237 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@45f6c803[dv_coll_shard4_replica_n6] main]
   [junit4]   2> 8238 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [dv_coll_shard4_replica_n6]  CLOSING SolrCore org.apache.solr.core.SolrCore@ee3913b
   [junit4]   2> 8238 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.dv_coll.shard4.replica_n6, tag=249794875
   [junit4]   2> 8238 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@31c87dbb: rootName = solr_65156, domain = solr.core.dv_coll.shard4.replica_n6, service url = null, agent id = null] for registry solr.core.dv_coll.shard4.replica_n6 / com.codahale.metrics.MetricRegistry@61829e8d
   [junit4]   2> 8240 INFO  (searcherExecutor-37-thread-1-processing-n:127.0.0.1:65156_solr x:dv_coll_shard4_replica_n6 s:shard4 c:dv_coll r:core_node8) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [dv_coll_shard4_replica_n6] Registered new searcher Searcher@45f6c803[dv_coll_shard4_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 8245 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard4.leader, tag=249794875
   [junit4]   2> 8246 INFO  (Thread-20) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore config update listener called for core dv_coll_shard2_replica_n2
   [junit4]   2> 8251 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DefaultSolrCoreState New IndexWriter is ready to be used.
   [junit4]   2> 8256 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@6dd8d4ea[dv_coll_shard3_replica_n4] main]
   [junit4]   2> 8256 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4]  CLOSING SolrCore org.apache.solr.core.SolrCore@276ed142
   [junit4]   2> 8256 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.dv_coll.shard3.replica_n4, tag=661573954
   [junit4]   2> 8256 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@1a69b5d3: rootName = solr_65157, domain = solr.core.dv_coll.shard3.replica_n4, service url = null, agent id = null] for registry solr.core.dv_coll.shard3.replica_n4 / com.codahale.metrics.MetricRegistry@540bf550
   [junit4]   2> 8258 INFO  (searcherExecutor-38-thread-1-processing-n:127.0.0.1:65157_solr x:dv_coll_shard3_replica_n4 s:shard3 c:dv_coll r:core_node7) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard3_replica_n4] Registered new searcher Searcher@6dd8d4ea[dv_coll_shard3_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 8262 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard3.leader, tag=661573954
   [junit4]   2> 8263 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore config update listener called for core dv_coll_shard1_replica_n1
   [junit4]   2> 8265 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore core reload dv_coll_shard1_replica_n1
   [junit4]   2> 8283 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 8283 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 8286 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@49ae5604[dv_coll_shard1_replica_n1] main]
   [junit4]   2> 8288 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 8288 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 8288 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 8289 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ZkIndexSchemaReader Current schema version 1 is already the latest
   [junit4]   2> 8290 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 8290 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1588529589278212096
   [junit4]   2> 8293 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DefaultSolrCoreState New IndexWriter is ready to be used.
   [junit4]   2> 8294 INFO  (searcherExecutor-39-thread-1-processing-n:127.0.0.1:65157_solr x:dv_coll_shard1_replica_n1 s:shard1 c:dv_coll r:core_node3) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@49ae5604[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 8294 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@704ee69b[dv_coll_shard1_replica_n1] main]
   [junit4]   2> 8295 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.ZkIndexSchemaReader Retrieved schema version 1 from ZooKeeper
   [junit4]   2> 8297 INFO  (searcherExecutor-39-thread-1-processing-n:127.0.0.1:65157_solr x:dv_coll_shard1_replica_n1 s:shard1 c:dv_coll r:core_node3) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@704ee69b[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 8299 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ManagedIndexSchema Waiting up to 599 secs for 3 replicas to apply schema update version 1 for collection dv_coll
   [junit4]   2> 8300 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 8305 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 8305 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.s.ZkIndexSchemaReader Finished refreshing schema in 10 ms
   [junit4]   2> 8306 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.3.0
   [junit4]   2> 8312 INFO  (qtp463211558-38) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.S.Request [dv_coll_shard3_replica_n4]  webapp=/solr path=/schema/zkversion params={refreshIfBelowVersion=1&wt=javabin&version=2} status=0 QTime=3
   [junit4]   2> 8312 INFO  (qtp655761501-26) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.S.Request [dv_coll_shard4_replica_n6]  webapp=/solr path=/schema/zkversion params={refreshIfBelowVersion=1&wt=javabin&version=2} status=0 QTime=1
   [junit4]   2> 8312 INFO  (qtp655761501-33) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.S.Request [dv_coll_shard2_replica_n2]  webapp=/solr path=/schema/zkversion params={refreshIfBelowVersion=1&wt=javabin&version=2} status=0 QTime=6
   [junit4]   2> 8317 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.s.ManagedIndexSchema Took 22.0ms for 3 replicas to apply schema update version 1 for collection dv_coll
   [junit4]   2> 8317 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.S.Request [dv_coll_shard1_replica_n1]  webapp=/solr path=/schema params={wt=javabin&version=2} status=0 QTime=374
   [junit4]   2> 8318 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1]  CLOSING SolrCore org.apache.solr.core.SolrCore@4023a5cb
   [junit4]   2> 8318 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.dv_coll.shard1.replica_n1, tag=1076078027
   [junit4]   2> 8318 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@534b9e78: rootName = solr_65157, domain = solr.core.dv_coll.shard1.replica_n1, service url = null, agent id = null] for registry solr.core.dv_coll.shard1.replica_n1 / com.codahale.metrics.MetricRegistry@50679c29
   [junit4]   2> 8324 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.IndexSchema [dv_coll_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 8335 INFO  (qtp463211558-37) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard1.leader, tag=1076078027
   [junit4]   2> 8337 INFO  (OverseerCollectionConfigSetProcessor-99283098800095238-127.0.0.1:65156_solr-n_0000000000) [n:127.0.0.1:65156_solr    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 8338 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 8338 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.CoreContainer Reloading SolrCore 'dv_coll_shard1_replica_n1' using configuration from collection dv_coll
   [junit4]   2> 8360 INFO  (TEST-DocValuesNotIndexedTest.testGroupingDVOnly-seed#[CDE0861F80EE206D]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testGroupingDVOnly
   [junit4]   2> 8378 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_65157.solr.core.dv_coll.shard1.replica_n1' (registry 'solr.core.dv_coll.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@55df545c
   [junit4]   2> 8378 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [[dv_coll_shard1_replica_n1] ] Opening new SolrCore at [/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/dv_coll_shard1_replica_n1], dataDir=[/Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001/tempDir-001/node2/./dv_coll_shard1_replica_n1/data/]
   [junit4]   2> 8434 INFO  (qtp655761501-34) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard2_replica_n2]  webapp=/solr path=/update params={update.distrib=TOLEADER&distrib.from=http://127.0.0.1:65157/solr/dv_coll_shard3_replica_n4/&wt=javabin&version=2}{deleteByQuery=*:* (-1588529589379923968)} 0 47
   [junit4]   2> 8435 INFO  (qtp655761501-28) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard4_replica_n6]  webapp=/solr path=/update params={update.distrib=TOLEADER&distrib.from=http://127.0.0.1:65157/solr/dv_coll_shard3_replica_n4/&wt=javabin&version=2}{deleteByQuery=*:* (-1588529589379923968)} 0 48
   [junit4]   2> 8436 INFO  (qtp463211558-36) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard1_replica_n1]  webapp=/solr path=/update params={update.distrib=TOLEADER&distrib.from=http://127.0.0.1:65157/solr/dv_coll_shard3_replica_n4/&wt=javabin&version=2}{deleteByQuery=*:* (-1588529589380972544)} 0 48
   [junit4]   2> 8437 INFO  (qtp463211558-25) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4]  webapp=/solr path=/update params={wt=javabin&version=2}{deleteByQuery=*:* (-1588529589375729664)} 0 73
   [junit4]   2> 8442 INFO  (qtp655761501-26) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1588529589437595648,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 8443 INFO  (qtp655761501-26) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 8442 INFO  (qtp463211558-38) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1588529589437595648,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 8443 INFO  (qtp463211558-38) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 8444 INFO  (qtp655761501-26) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@4e2489dd[dv_coll_shard2_replica_n2] main]
   [junit4]   2> 8445 INFO  (qtp463211558-39) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1588529589440741376,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 8445 INFO  (qtp463211558-39) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 8446 INFO  (qtp655761501-26) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 8446 INFO  (qtp463211558-38) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 8446 INFO  (qtp463211558-38) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard1_replica_n1]  webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:65157/solr/dv_coll_shard1_replica_n1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 4
   [junit4]   2> 8447 INFO  (qtp463211558-39) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 8447 INFO  (qtp463211558-39) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4]  webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:65157/solr/dv_coll_shard1_replica_n1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 2
   [junit4]   2> 8448 INFO  (searcherExecutor-19-thread-1-processing-n:127.0.0.1:65156_solr x:dv_coll_shard2_replica_n2 s:shard2 c:dv_coll r:core_node5) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [dv_coll_shard2_replica_n2] Registered new searcher Searcher@4e2489dd[dv_coll_shard2_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 8448 INFO  (qtp655761501-26) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard2_replica_n2]  webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:65157/solr/dv_coll_shard1_replica_n1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 6
   [junit4]   2> 8448 INFO  (qtp655761501-33) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 start commit{_version_=1588529589443887104,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 8449 INFO  (qtp655761501-33) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 8453 INFO  (qtp655761501-33) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.DirectUpdateHandler2 end_commit_flush
   [junit4]   2> 8453 INFO  (qtp655761501-33) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard4_replica_n6]  webapp=/solr path=/update params={update.distrib=FROMLEADER&waitSearcher=true&openSearcher=true&commit=true&softCommit=false&distrib.from=http://127.0.0.1:65157/solr/dv_coll_shard1_replica_n1/&commit_end_point=true&wt=javabin&version=2&expungeDeletes=false}{commit=} 0 5
   [junit4]   2> 8454 INFO  (qtp463211558-35) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard1_replica_n1]  webapp=/solr path=/update params={_stateVer_=dv_coll:4&waitSearcher=true&commit=true&softCommit=false&wt=javabin&version=2}{commit=} 0 15
   [junit4]   2> 8466 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 8466 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 8477 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@3b715147[dv_coll_shard1_replica_n1] main]
   [junit4]   2> 8487 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 8488 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 8488 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema
   [junit4]   2> 8489 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.s.ZkIndexSchemaReader Current schema version 1 is already the latest
   [junit4]   2> 8490 INFO  (Thread-21) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 8492 INFO  (searcherExecutor-50-thread-1-processing-n:127.0.0.1:65157_solr x:dv_coll_shard3_replica_n4 s:shard3 c:dv_coll r:core_node7) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.c.SolrCore [dv_coll_shard1_replica_n1] Registered new searcher Searcher@3b715147[dv_coll_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 8528 INFO  (qtp655761501-30) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard2_replica_n2]  webapp=/solr path=/update params={_stateVer_=dv_coll:4&wt=javabin&version=2}{add=[0 (1588529589460664320), 4 (1588529589515190272), 10005 (1588529589516238848), 12 (1588529589516238849), 15 (1588529589520433152), 16 (1588529589521481728), 20 (1588529589521481729), 22 (1588529589522530304), 34 (1588529589523578880), 35 (1588529589524627456), ... (16 adds)]} 0 65
   [junit4]   2> 8528 INFO  (qtp655761501-29) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard4_replica_n6]  webapp=/solr path=/update params={_stateVer_=dv_coll:4&wt=javabin&version=2}{add=[10000 (1588529589460664320), 5 (1588529589517287424), 17 (1588529589520433152), 19 (1588529589521481728), 21 (1588529589523578880), 23 (1588529589524627456), 25 (1588529589524627457), 29 (1588529589525676032), 10030 (1588529589525676033), 33 (1588529589525676034), ... (13 adds)]} 0 65
   [junit4]   2> 8533 INFO  (qtp463211558-40) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.u.p.LogUpdateProcessorFactory [dv_coll_shard3_replica_n4]  webapp=/solr path=/update params={_stateVer_=dv_coll:4&wt=javabin&version=2}{add=[2 (1588529589460664320), 3 (1588529589517287424), 6 (1588529589518336000), 7 (15885295895183360

[...truncated too long message...]

re org.apache.solr.core.SolrCore@65c54024
   [junit4]   2> 9848 INFO  (zkCallback-19-thread-3-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (0)
   [junit4]   2> 9848 INFO  (zkCallback-17-thread-2-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (0)
   [junit4]   2> 9849 INFO  (zkCallback-19-thread-2-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [0])
   [junit4]   2> 9849 INFO  (zkCallback-17-thread-1-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [0])
   [junit4]   2> 9849 INFO  (zkCallback-17-thread-3-processing-n:127.0.0.1:65157_solr) [n:127.0.0.1:65157_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [0])
   [junit4]   2> 9849 INFO  (zkCallback-19-thread-1-processing-n:127.0.0.1:65156_solr) [n:127.0.0.1:65156_solr    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/dv_coll/state.json] for collection [dv_coll] has occurred - updating... (live nodes size: [0])
   [junit4]   2> 9852 INFO  (jetty-closer-2-thread-2) [    ] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.cluster, tag=null
   [junit4]   2> 9853 INFO  (coreCloseExecutor-56-thread-1) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.c.SolrCore [dv_coll_shard4_replica_n6]  CLOSING SolrCore org.apache.solr.core.SolrCore@1d56e45f
   [junit4]   2> 9853 INFO  (coreCloseExecutor-56-thread-1) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.dv_coll.shard4.replica_n6, tag=492233823
   [junit4]   2> 9853 INFO  (coreCloseExecutor-56-thread-1) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@647c186f: rootName = solr_65156, domain = solr.core.dv_coll.shard4.replica_n6, service url = null, agent id = null] for registry solr.core.dv_coll.shard4.replica_n6 / com.codahale.metrics.MetricRegistry@61829e8d
   [junit4]   2> 9854 INFO  (coreCloseExecutor-56-thread-2) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.c.SolrCore [dv_coll_shard2_replica_n2]  CLOSING SolrCore org.apache.solr.core.SolrCore@57d6dea6
   [junit4]   2> 9856 INFO  (coreCloseExecutor-55-thread-1) [n:127.0.0.1:65157_solr c:dv_coll s:shard3 r:core_node7 x:dv_coll_shard3_replica_n4] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard3.leader, tag=174230049
   [junit4]   2> 9856 INFO  (coreCloseExecutor-55-thread-2) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.dv_coll.shard1.replica_n1, tag=1707425828
   [junit4]   2> 9857 INFO  (coreCloseExecutor-55-thread-2) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@3037c5fe: rootName = solr_65157, domain = solr.core.dv_coll.shard1.replica_n1, service url = null, agent id = null] for registry solr.core.dv_coll.shard1.replica_n1 / com.codahale.metrics.MetricRegistry@50679c29
   [junit4]   2> 9863 INFO  (coreCloseExecutor-56-thread-1) [n:127.0.0.1:65156_solr c:dv_coll s:shard4 r:core_node8 x:dv_coll_shard4_replica_n6] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard4.leader, tag=492233823
   [junit4]   2> 9863 INFO  (coreCloseExecutor-56-thread-2) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.core.dv_coll.shard2.replica_n2, tag=1473699494
   [junit4]   2> 9863 INFO  (coreCloseExecutor-56-thread-2) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.m.r.SolrJmxReporter Closing reporter [org.apache.solr.metrics.reporters.SolrJmxReporter@d502ad7: rootName = solr_65156, domain = solr.core.dv_coll.shard2.replica_n2, service url = null, agent id = null] for registry solr.core.dv_coll.shard2.replica_n2 / com.codahale.metrics.MetricRegistry@40ec95a9
   [junit4]   2> 9875 INFO  (coreCloseExecutor-55-thread-2) [n:127.0.0.1:65157_solr c:dv_coll s:shard1 r:core_node3 x:dv_coll_shard1_replica_n1] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard1.leader, tag=1707425828
   [junit4]   2> 9881 INFO  (coreCloseExecutor-56-thread-2) [n:127.0.0.1:65156_solr c:dv_coll s:shard2 r:core_node5 x:dv_coll_shard2_replica_n2] o.a.s.m.SolrMetricManager Closing metric reporters for registry=solr.collection.dv_coll.shard2.leader, tag=1473699494
   [junit4]   2> 9904 INFO  (jetty-closer-2-thread-1) [    ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@7ae8e815{/solr,null,UNAVAILABLE}
   [junit4]   2> 9910 INFO  (jetty-closer-2-thread-2) [    ] o.a.s.c.Overseer Overseer (id=99283098800095238-127.0.0.1:65156_solr-n_0000000000) closing
   [junit4]   2> 9910 INFO  (OverseerStateUpdate-99283098800095238-127.0.0.1:65156_solr-n_0000000000) [n:127.0.0.1:65156_solr    ] o.a.s.c.Overseer Overseer Loop exiting : 127.0.0.1:65156_solr
   [junit4]   2> 9910 WARN  (OverseerAutoScalingTriggerThread-99283098800095238-127.0.0.1:65156_solr-n_0000000000) [n:127.0.0.1:65156_solr    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread woken up but we are closed, exiting.
   [junit4]   2> 9913 INFO  (jetty-closer-2-thread-2) [    ] o.e.j.s.h.ContextHandler Stopped o.e.j.s.ServletContextHandler@2694c442{/solr,null,UNAVAILABLE}
   [junit4]   2> 9914 ERROR (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 9915 INFO  (SUITE-DocValuesNotIndexedTest-seed#[CDE0861F80EE206D]-worker) [    ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:65145 65145
   [junit4]   2> 9934 INFO  (Thread-0) [    ] o.a.s.c.ZkTestServer connecting to 127.0.0.1:65145 65145
   [junit4]   2> 10109 WARN  (Thread-0) [    ] o.a.s.c.ZkTestServer Watch limit violations: 
   [junit4]   2> Maximum concurrent create/delete watches above limit:
   [junit4]   2> 
   [junit4]   2> 	8	/solr/configs/conf1/managed-schema
   [junit4]   2> 	4	/solr/configs/conf1
   [junit4]   2> 	3	/solr/aliases.json
   [junit4]   2> 	3	/solr/clusterprops.json
   [junit4]   2> 	2	/solr/security.json
   [junit4]   2> 
   [junit4]   2> Maximum concurrent data watches above limit:
   [junit4]   2> 
   [junit4]   2> 	12	/solr/collections/dv_coll/state.json
   [junit4]   2> 	3	/solr/clusterstate.json
   [junit4]   2> 	2	/solr/configs/conf1/managed-schema
   [junit4]   2> 
   [junit4]   2> Maximum concurrent children watches above limit:
   [junit4]   2> 
   [junit4]   2> 	3	/solr/live_nodes
   [junit4]   2> 	3	/solr/collections
   [junit4]   2> 
   [junit4]   2> NOTE: leaving temporary files on disk at: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.DocValuesNotIndexedTest_CDE0861F80EE206D-001
   [junit4]   2> Jan 03, 2018 12:37:51 AM com.carrotsearch.randomizedtesting.ThreadLeakControl checkThreadLeaks
   [junit4]   2> WARNING: Will linger awaiting termination of 1 leaked thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene70): {id=FST50}, docValues:{floatField=DocValuesFormat(name=Lucene70), doubleField=DocValuesFormat(name=Lucene70), intFieldMulti=DocValuesFormat(name=Lucene70), boolFieldMulti=DocValuesFormat(name=Lucene70), longFieldMulti=DocValuesFormat(name=Memory), floatGSL=DocValuesFormat(name=Asserting), intField=DocValuesFormat(name=Lucene70), boolField=DocValuesFormat(name=Asserting), stringFieldMulti=DocValuesFormat(name=Asserting), dateFieldMulti=DocValuesFormat(name=Memory), longGSF=DocValuesFormat(name=Lucene70), doubleFieldMulti=DocValuesFormat(name=Asserting), stringGSL=DocValuesFormat(name=Memory), doubleGSF=DocValuesFormat(name=Lucene70), doubleGSL=DocValuesFormat(name=Memory), stringGSF=DocValuesFormat(name=Lucene70), intGSL=DocValuesFormat(name=Lucene70), longField=DocValuesFormat(name=Lucene70), dateField=DocValuesFormat(name=Asserting), dateGSF=DocValuesFormat(name=Asserting), boolGSF=DocValuesFormat(name=Asserting), longGSL=DocValuesFormat(name=Asserting), boolGSL=DocValuesFormat(name=Lucene70), floatFieldMulti=DocValuesFormat(name=Lucene70), floatGSF=DocValuesFormat(name=Lucene70), dateGSL=DocValuesFormat(name=Lucene70), _version_=DocValuesFormat(name=Lucene70), intGSF=DocValuesFormat(name=Memory), stringField=DocValuesFormat(name=Lucene70)}, maxPointsInLeafNode=406, maxMBSortInHeap=6.717276710907141, sim=RandomSimilarity(queryNorm=true): {}, locale=en-ZW, timezone=America/Buenos_Aires
   [junit4]   2> NOTE: Mac OS X 10.11.6 x86_64/Oracle Corporation 9 (64-bit)/cpus=3,threads=1,free=154356752,total=228589568
   [junit4]   2> NOTE: All tests run in this JVM: [DocValuesNotIndexedTest]
   [junit4] Completed [1/764 (1!)] on J1 in 10.96s, 4 tests, 1 failure <<< FAILURES!

[...truncated 43409 lines...]

[JENKINS] Lucene-Solr-7.x-MacOSX (64bit/jdk-9) - Build # 375 - Still Unstable!

Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: https://jenkins.thetaphi.de/job/Lucene-Solr-7.x-MacOSX/375/
Java: 64bit/jdk-9 -XX:-UseCompressedOops -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  org.apache.solr.cloud.autoscaling.sim.TestLargeCluster.testAddNode

Error Message:
did not finish processing changes

Stack Trace:
java.lang.AssertionError: did not finish processing changes
	at __randomizedtesting.SeedInfo.seed([D6A597F2D5DB4579:714A8A511A96CA61]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.junit.Assert.assertTrue(Assert.java:43)
	at org.apache.solr.cloud.autoscaling.sim.TestLargeCluster.testAddNode(TestLargeCluster.java:297)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.base/java.lang.reflect.Method.invoke(Method.java:564)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.base/java.lang.Thread.run(Thread.java:844)




Build Log:
[...truncated 1735 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/core/test/temp/junit4-J0-20180102_132303_03611496602337465376302.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/core/test/temp/junit4-J1-20180102_132303_0366330421227022086400.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 288 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/test-framework/test/temp/junit4-J0-20180102_132958_38814652159752657278286.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/test-framework/test/temp/junit4-J1-20180102_132958_3889625478022658073674.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 1057 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/common/test/temp/junit4-J0-20180102_133108_4128397174531284266604.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/common/test/temp/junit4-J1-20180102_133108_412801340335916895435.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 229 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/icu/test/temp/junit4-J0-20180102_133325_74913497201679895896738.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/icu/test/temp/junit4-J1-20180102_133325_749176715766720666896.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 249 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/kuromoji/test/temp/junit4-J0-20180102_133335_7864104111128380125865.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/kuromoji/test/temp/junit4-J1-20180102_133335_7868586923737448302823.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 157 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/morfologik/test/temp/junit4-J1-20180102_133415_5578317726383767107241.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/morfologik/test/temp/junit4-J0-20180102_133415_5574606348572104909850.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 161 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/opennlp/test/temp/junit4-J0-20180102_133420_1019934646160142875281.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/opennlp/test/temp/junit4-J1-20180102_133420_1016492993032172193088.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 168 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/phonetic/test/temp/junit4-J1-20180102_133423_6187170428273383068427.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/phonetic/test/temp/junit4-J0-20180102_133423_61818342691189866070236.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 151 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/smartcn/test/temp/junit4-J1-20180102_133433_57814650207479570606687.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/smartcn/test/temp/junit4-J0-20180102_133433_578539391646284818223.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 159 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/stempel/test/temp/junit4-J1-20180102_133444_25812609420741927535478.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/stempel/test/temp/junit4-J0-20180102_133444_2582841579067839005817.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 167 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/analysis/uima/test/temp/junit4-J0-20180102_133447_82213360821526107013281.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 213 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/backward-codecs/test/temp/junit4-J0-20180102_133501_13312875706408711790375.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 9 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/backward-codecs/test/temp/junit4-J1-20180102_133501_1335348354729007203027.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 1317 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/benchmark/test/temp/junit4-J1-20180102_133630_41618198352252084622775.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/benchmark/test/temp/junit4-J0-20180102_133630_4165778700642051297058.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 231 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/classification/test/temp/junit4-J0-20180102_133640_9288580964719411831246.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/classification/test/temp/junit4-J1-20180102_133640_92815621076257875867181.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 259 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/codecs/test/temp/junit4-J1-20180102_133658_13115445130771222816124.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/codecs/test/temp/junit4-J0-20180102_133658_13113385439779281522208.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 212 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/demo/test/temp/junit4-J0-20180102_133846_75610842728736879579078.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/demo/test/temp/junit4-J1-20180102_133846_7566852527371778548602.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 164 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/expressions/test/temp/junit4-J0-20180102_133851_7142625060356208061226.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 9 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/expressions/test/temp/junit4-J1-20180102_133851_71513185391310525039143.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 229 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/facet/test/temp/junit4-J0-20180102_133859_5206983539925313179135.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/facet/test/temp/junit4-J1-20180102_133859_52014775062758042089897.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 182 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/grouping/test/temp/junit4-J1-20180102_133933_0468525913922019679343.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/grouping/test/temp/junit4-J0-20180102_133933_04612912247919770969302.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 243 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/highlighter/test/temp/junit4-J0-20180102_133940_7185483481265108436805.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/highlighter/test/temp/junit4-J1-20180102_133940_71817511991630167059263.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 163 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/join/test/temp/junit4-J1-20180102_134000_29610344807491044630798.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/join/test/temp/junit4-J0-20180102_134000_29618438242117117765588.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 147 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/memory/test/temp/junit4-J0-20180102_134016_0371104566833584686322.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/memory/test/temp/junit4-J1-20180102_134016_03716375425050743683271.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 187 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/misc/test/temp/junit4-J0-20180102_134023_00212596081675232700888.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/misc/test/temp/junit4-J1-20180102_134023_00218392300686704997363.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 329 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/queries/test/temp/junit4-J0-20180102_134040_01317353993532963494656.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/queries/test/temp/junit4-J1-20180102_134040_013648986045860022378.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 224 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/queryparser/test/temp/junit4-J1-20180102_134050_938544128589304484985.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/queryparser/test/temp/junit4-J0-20180102_134050_9389603302092081744415.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 207 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/replicator/test/temp/junit4-J1-20180102_134100_6534531462394154802429.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/replicator/test/temp/junit4-J0-20180102_134100_6536851521224079581250.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 216 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/sandbox/test/temp/junit4-J1-20180102_134113_8997329382087892881945.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 7 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/sandbox/test/temp/junit4-J0-20180102_134113_8992380989082383498024.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 317 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial-extras/test/temp/junit4-J1-20180102_134147_08916229481512868295063.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial-extras/test/temp/junit4-J0-20180102_134147_08917656654467019358947.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 146 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial/test/temp/junit4-J0-20180102_134202_61811140234424310835849.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 177 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial3d/test/temp/junit4-J1-20180102_134204_8496200659989690035342.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 5 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/spatial3d/test/temp/junit4-J0-20180102_134204_84916612740132525519051.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 250 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/suggest/test/temp/junit4-J1-20180102_134220_3046624456237972283264.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/lucene/build/suggest/test/temp/junit4-J0-20180102_134220_3041509444876976012819.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 2426 lines...]
   [junit4] Suite: org.apache.solr.cloud.autoscaling.sim.TestLargeCluster
   [junit4]   2> 1426861 INFO  (SUITE-TestLargeCluster-seed#[D6A597F2D5DB4579]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> Creating dataDir: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-core/test/J1/temp/solr.cloud.autoscaling.sim.TestLargeCluster_D6A597F2D5DB4579-001/init-core-data-001
   [junit4]   2> 1426862 WARN  (SUITE-TestLargeCluster-seed#[D6A597F2D5DB4579]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=1 numCloses=1
   [junit4]   2> 1426862 INFO  (SUITE-TestLargeCluster-seed#[D6A597F2D5DB4579]-worker) [    ] o.a.s.SolrTestCaseJ4 Using TrieFields (NUMERIC_POINTS_SYSPROP=false) w/NUMERIC_DOCVALUES_SYSPROP=false
   [junit4]   2> 1426863 INFO  (SUITE-TestLargeCluster-seed#[D6A597F2D5DB4579]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason="", value=0.0/0.0, ssl=0.0/0.0, clientAuth=0.0/0.0) w/ MAC_OS_X supressed clientAuth
   [junit4]   2> 1426865 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Adding .autoAddReplicas trigger
   [junit4]   2> 1426865 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 0
   [junit4]   2> 1426865 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeLostTrigger Initial livenodes: [127.0.0.1:10022_solr, 127.0.0.1:10018_solr, 127.0.0.1:10025_solr, 127.0.0.1:10017_solr, 127.0.0.1:10026_solr, 127.0.0.1:10012_solr, 127.0.0.1:10020_solr, 127.0.0.1:10028_solr, 127.0.0.1:10023_solr, 127.0.0.1:10015_solr, 127.0.0.1:10016_solr, 127.0.0.1:10013_solr, 127.0.0.1:10021_solr, 127.0.0.1:10029_solr, 127.0.0.1:10030_solr, 127.0.0.1:10014_solr, 127.0.0.1:10019_solr, 127.0.0.1:10024_solr, 127.0.0.1:10027_solr]
   [junit4]   2> 1426865 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 0, lastZnodeVersion -1
   [junit4]   2> 1426865 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 0
   [junit4]   2> 1426865 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 1426865 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 0, lastZnodeVersion 0
   [junit4]   2> 1426866 DEBUG (ScheduledTrigger-5341-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 40
   [junit4]   2> 1426869 INFO  (TEST-TestLargeCluster.testAddNode-seed#[D6A597F2D5DB4579]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testAddNode
   [junit4]   2> 1426869 DEBUG (TEST-TestLargeCluster.testAddNode-seed#[D6A597F2D5DB4579]) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 1
   [junit4]   2> 1426869 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 1
   [junit4]   2> 1426870 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeLost markers
   [junit4]   2> 1426870 INFO  (TEST-TestLargeCluster.testAddNode-seed#[D6A597F2D5DB4579]) [    ] o.a.s.c.a.s.SimCloudManager === Restarting OverseerTriggerThread and clearing object cache...
   [junit4]   2> 1426870 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 1426870 WARN  (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Interrupted
   [junit4]   2> java.lang.InterruptedException
   [junit4]   2> 	at java.base/java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1247)
   [junit4]   2> 	at java.base/java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:317)
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:154)
   [junit4]   2> 	at java.base/java.lang.Thread.run(Thread.java:844)
   [junit4]   2> 1426870 DEBUG (TEST-TestLargeCluster.testAddNode-seed#[D6A597F2D5DB4579]) [    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread has been closed explicitly
   [junit4]   2> 1426871 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Adding .autoAddReplicas trigger
   [junit4]   2> 1426871 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 2
   [junit4]   2> 1426871 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeLostTrigger Initial livenodes: [127.0.0.1:10082_solr, 127.0.0.1:10018_solr, 127.0.0.1:10049_solr, 127.0.0.1:10104_solr, 127.0.0.1:10020_solr, 127.0.0.1:10065_solr, 127.0.0.1:10021_solr, 127.0.0.1:10066_solr, 127.0.0.1:10035_solr, 127.0.0.1:10051_solr, 127.0.0.1:10019_solr, 127.0.0.1:10096_solr, 127.0.0.1:10022_solr, 127.0.0.1:10033_solr, 127.0.0.1:10105_solr, 127.0.0.1:10017_solr, 127.0.0.1:10047_solr, 127.0.0.1:10102_solr, 127.0.0.1:10095_solr, 127.0.0.1:10081_solr, 127.0.0.1:10084_solr, 127.0.0.1:10036_solr, 127.0.0.1:10052_solr, 127.0.0.1:10079_solr, 127.0.0.1:10098_solr, 127.0.0.1:10054_solr, 127.0.0.1:10099_solr, 127.0.0.1:10101_solr, 127.0.0.1:10071_solr, 127.0.0.1:10023_solr, 127.0.0.1:10076_solr, 127.0.0.1:10107_solr, 127.0.0.1:10055_solr, 127.0.0.1:10016_solr, 127.0.0.1:10070_solr, 127.0.0.1:10077_solr, 127.0.0.1:10038_solr, 127.0.0.1:10032_solr, 127.0.0.1:10108_solr, 127.0.0.1:10093_solr, 127.0.0.1:10061_solr, 127.0.0.1:10067_solr, 127.0.0.1:10086_solr, 127.0.0.1:10089_solr, 127.0.0.1:10045_solr, 127.0.0.1:10042_solr, 127.0.0.1:10029_solr, 127.0.0.1:10083_solr, 127.0.0.1:10080_solr, 127.0.0.1:10048_solr, 127.0.0.1:10064_solr, 127.0.0.1:10111_solr, 127.0.0.1:10043_solr, 127.0.0.1:10057_solr, 127.0.0.1:10074_solr, 127.0.0.1:10026_solr, 127.0.0.1:10012_solr, 127.0.0.1:10060_solr, 127.0.0.1:10073_solr, 127.0.0.1:10058_solr, 127.0.0.1:10013_solr, 127.0.0.1:10090_solr, 127.0.0.1:10088_solr, 127.0.0.1:10027_solr, 127.0.0.1:10039_solr, 127.0.0.1:10025_solr, 127.0.0.1:10044_solr, 127.0.0.1:10028_solr, 127.0.0.1:10092_solr, 127.0.0.1:10041_solr, 127.0.0.1:10087_solr, 127.0.0.1:10030_solr, 127.0.0.1:10014_solr, 127.0.0.1:10040_solr, 127.0.0.1:10085_solr, 127.0.0.1:10046_solr, 127.0.0.1:10068_solr, 127.0.0.1:10063_solr, 127.0.0.1:10015_solr, 127.0.0.1:10069_solr, 127.0.0.1:10110_solr, 127.0.0.1:10062_solr, 127.0.0.1:10024_solr, 127.0.0.1:10050_solr, 127.0.0.1:10075_solr, 127.0.0.1:10078_solr, 127.0.0.1:10097_solr, 127.0.0.1:10031_solr, 127.0.0.1:10034_solr, 127.0.0.1:10091_solr, 127.0.0.1:10094_solr, 127.0.0.1:10106_solr, 127.0.0.1:10100_solr, 127.0.0.1:10037_solr, 127.0.0.1:10109_solr, 127.0.0.1:10053_solr, 127.0.0.1:10059_solr, 127.0.0.1:10103_solr, 127.0.0.1:10056_solr, 127.0.0.1:10072_solr]
   [junit4]   2> 1426871 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 2, lastZnodeVersion -1
   [junit4]   2> 1426871 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 2
   [junit4]   2> 1426871 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 1426871 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 2, lastZnodeVersion 2
   [junit4]   2> 1426872 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1426898 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1426921 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1426945 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1426967 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1426981 DEBUG (simCloudManagerPool-5343-thread-1) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 1426981 DEBUG (simCloudManagerPool-5343-thread-1) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 3
   [junit4]   2> 1426981 DEBUG (simCloudManagerPool-5343-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Initial livenodes: [127.0.0.1:10082_solr, 127.0.0.1:10018_solr, 127.0.0.1:10049_solr, 127.0.0.1:10104_solr, 127.0.0.1:10020_solr, 127.0.0.1:10065_solr, 127.0.0.1:10021_solr, 127.0.0.1:10066_solr, 127.0.0.1:10035_solr, 127.0.0.1:10051_solr, 127.0.0.1:10019_solr, 127.0.0.1:10096_solr, 127.0.0.1:10022_solr, 127.0.0.1:10033_solr, 127.0.0.1:10105_solr, 127.0.0.1:10017_solr, 127.0.0.1:10047_solr, 127.0.0.1:10102_solr, 127.0.0.1:10095_solr, 127.0.0.1:10081_solr, 127.0.0.1:10084_solr, 127.0.0.1:10036_solr, 127.0.0.1:10052_solr, 127.0.0.1:10079_solr, 127.0.0.1:10098_solr, 127.0.0.1:10054_solr, 127.0.0.1:10099_solr, 127.0.0.1:10101_solr, 127.0.0.1:10071_solr, 127.0.0.1:10023_solr, 127.0.0.1:10076_solr, 127.0.0.1:10107_solr, 127.0.0.1:10055_solr, 127.0.0.1:10016_solr, 127.0.0.1:10070_solr, 127.0.0.1:10077_solr, 127.0.0.1:10038_solr, 127.0.0.1:10032_solr, 127.0.0.1:10108_solr, 127.0.0.1:10093_solr, 127.0.0.1:10061_solr, 127.0.0.1:10067_solr, 127.0.0.1:10086_solr, 127.0.0.1:10089_solr, 127.0.0.1:10045_solr, 127.0.0.1:10042_solr, 127.0.0.1:10029_solr, 127.0.0.1:10083_solr, 127.0.0.1:10080_solr, 127.0.0.1:10048_solr, 127.0.0.1:10064_solr, 127.0.0.1:10111_solr, 127.0.0.1:10043_solr, 127.0.0.1:10057_solr, 127.0.0.1:10074_solr, 127.0.0.1:10026_solr, 127.0.0.1:10012_solr, 127.0.0.1:10060_solr, 127.0.0.1:10073_solr, 127.0.0.1:10058_solr, 127.0.0.1:10013_solr, 127.0.0.1:10090_solr, 127.0.0.1:10088_solr, 127.0.0.1:10027_solr, 127.0.0.1:10039_solr, 127.0.0.1:10025_solr, 127.0.0.1:10044_solr, 127.0.0.1:10028_solr, 127.0.0.1:10092_solr, 127.0.0.1:10041_solr, 127.0.0.1:10087_solr, 127.0.0.1:10030_solr, 127.0.0.1:10014_solr, 127.0.0.1:10040_solr, 127.0.0.1:10085_solr, 127.0.0.1:10046_solr, 127.0.0.1:10068_solr, 127.0.0.1:10063_solr, 127.0.0.1:10015_solr, 127.0.0.1:10069_solr, 127.0.0.1:10110_solr, 127.0.0.1:10062_solr, 127.0.0.1:10024_solr, 127.0.0.1:10050_solr, 127.0.0.1:10075_solr, 127.0.0.1:10078_solr, 127.0.0.1:10097_solr, 127.0.0.1:10031_solr, 127.0.0.1:10034_solr, 127.0.0.1:10091_solr, 127.0.0.1:10094_solr, 127.0.0.1:10106_solr, 127.0.0.1:10100_solr, 127.0.0.1:10037_solr, 127.0.0.1:10109_solr, 127.0.0.1:10053_solr, 127.0.0.1:10059_solr, 127.0.0.1:10103_solr, 127.0.0.1:10056_solr, 127.0.0.1:10072_solr]
   [junit4]   2> 1426982 DEBUG (simCloudManagerPool-5343-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Initial livenodes: [127.0.0.1:10082_solr, 127.0.0.1:10018_solr, 127.0.0.1:10049_solr, 127.0.0.1:10104_solr, 127.0.0.1:10020_solr, 127.0.0.1:10065_solr, 127.0.0.1:10021_solr, 127.0.0.1:10066_solr, 127.0.0.1:10035_solr, 127.0.0.1:10051_solr, 127.0.0.1:10019_solr, 127.0.0.1:10096_solr, 127.0.0.1:10022_solr, 127.0.0.1:10033_solr, 127.0.0.1:10105_solr, 127.0.0.1:10017_solr, 127.0.0.1:10047_solr, 127.0.0.1:10102_solr, 127.0.0.1:10095_solr, 127.0.0.1:10081_solr, 127.0.0.1:10084_solr, 127.0.0.1:10036_solr, 127.0.0.1:10052_solr, 127.0.0.1:10079_solr, 127.0.0.1:10098_solr, 127.0.0.1:10054_solr, 127.0.0.1:10099_solr, 127.0.0.1:10101_solr, 127.0.0.1:10071_solr, 127.0.0.1:10023_solr, 127.0.0.1:10076_solr, 127.0.0.1:10107_solr, 127.0.0.1:10055_solr, 127.0.0.1:10016_solr, 127.0.0.1:10070_solr, 127.0.0.1:10077_solr, 127.0.0.1:10038_solr, 127.0.0.1:10032_solr, 127.0.0.1:10108_solr, 127.0.0.1:10093_solr, 127.0.0.1:10061_solr, 127.0.0.1:10067_solr, 127.0.0.1:10086_solr, 127.0.0.1:10089_solr, 127.0.0.1:10045_solr, 127.0.0.1:10042_solr, 127.0.0.1:10029_solr, 127.0.0.1:10083_solr, 127.0.0.1:10080_solr, 127.0.0.1:10048_solr, 127.0.0.1:10064_solr, 127.0.0.1:10111_solr, 127.0.0.1:10043_solr, 127.0.0.1:10057_solr, 127.0.0.1:10074_solr, 127.0.0.1:10026_solr, 127.0.0.1:10012_solr, 127.0.0.1:10060_solr, 127.0.0.1:10073_solr, 127.0.0.1:10058_solr, 127.0.0.1:10013_solr, 127.0.0.1:10090_solr, 127.0.0.1:10088_solr, 127.0.0.1:10027_solr, 127.0.0.1:10039_solr, 127.0.0.1:10025_solr, 127.0.0.1:10044_solr, 127.0.0.1:10028_solr, 127.0.0.1:10092_solr, 127.0.0.1:10041_solr, 127.0.0.1:10087_solr, 127.0.0.1:10030_solr, 127.0.0.1:10014_solr, 127.0.0.1:10040_solr, 127.0.0.1:10085_solr, 127.0.0.1:10046_solr, 127.0.0.1:10068_solr, 127.0.0.1:10063_solr, 127.0.0.1:10015_solr, 127.0.0.1:10069_solr, 127.0.0.1:10110_solr, 127.0.0.1:10062_solr, 127.0.0.1:10024_solr, 127.0.0.1:10050_solr, 127.0.0.1:10075_solr, 127.0.0.1:10078_solr, 127.0.0.1:10097_solr, 127.0.0.1:10031_solr, 127.0.0.1:10034_solr, 127.0.0.1:10091_solr, 127.0.0.1:10094_solr, 127.0.0.1:10106_solr, 127.0.0.1:10100_solr, 127.0.0.1:10037_solr, 127.0.0.1:10109_solr, 127.0.0.1:10053_solr, 127.0.0.1:10059_solr, 127.0.0.1:10103_solr, 127.0.0.1:10056_solr, 127.0.0.1:10072_solr]
   [junit4]   2> 1426982 DEBUG (simCloudManagerPool-5343-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger NodeAddedTrigger node_added_trigger instantiated with properties: {event=nodeAdded, waitFor=5, actions=[{name=compute, class=org.apache.solr.cloud.autoscaling.ComputePlanAction}, {name=execute, class=org.apache.solr.cloud.autoscaling.ExecutePlanAction}], enabled=true}
   [junit4]   2> 1426982 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 3
   [junit4]   2> 1426982 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 3, lastZnodeVersion 3
   [junit4]   2> 1426982 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1426982 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1426991 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427003 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427004 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427012 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427024 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427024 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427041 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427045 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427045 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427069 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427069 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427070 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427090 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427090 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427090 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427110 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427110 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427111 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427137 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427137 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427137 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427158 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427158 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427158 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427179 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427179 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427179 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427199 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427199 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427199 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427219 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427220 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427219 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427240 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427240 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427240 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427263 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427263 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427263 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427286 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427286 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427286 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427308 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427308 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427308 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427331 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427331 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427331 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427352 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427352 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427352 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427372 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427372 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427372 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427393 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427393 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427393 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427415 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427415 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427415 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427440 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427440 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427440 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427464 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427464 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427464 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427486 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427486 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427486 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427508 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427508 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427508 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427531 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427531 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427531 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427555 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427555 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427555 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427580 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427580 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427581 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427604 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427604 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427604 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427628 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427628 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427628 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427655 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427656 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427656 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427688 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427688 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427688 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427712 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427712 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427712 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427744 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427744 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427744 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427765 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427765 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427766 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427791 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427791 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427791 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427812 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427812 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427812 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427833 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427834 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427834 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427856 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427856 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427856 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427876 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427876 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427877 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427898 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427898 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427898 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427927 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427927 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427927 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427948 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427948 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427948 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427976 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427976 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427976 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1427997 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1427997 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1427997 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428018 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428018 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428018 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428040 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428040 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428040 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428061 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428061 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428061 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428081 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428082 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428082 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428104 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428105 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428105 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428125 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428125 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428125 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428148 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428149 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428149 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428171 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428171 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428171 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428195 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428195 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428195 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428217 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428217 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428217 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428238 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428239 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428239 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428264 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428264 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428264 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428290 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428290 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428290 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428316 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428316 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428316 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428342 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428342 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428342 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428370 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428370 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428370 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428396 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428396 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428396 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428423 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428423 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428423 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428448 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428448 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428448 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428473 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428473 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428473 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428502 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428502 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428502 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428529 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428529 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428529 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428556 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428556 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428556 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428577 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428577 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428577 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428606 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428606 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428607 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428636 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428636 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428637 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428663 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428664 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428664 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428684 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428684 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428684 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428714 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428714 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428714 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428744 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428745 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428745 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428768 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428768 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428768 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428791 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428791 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428791 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428820 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428820 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428820 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428847 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428847 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428847 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428873 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428873 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428873 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428900 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428900 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428900 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428928 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428928 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428928 DEBUG (ScheduledTrigger-5344-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428955 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428955 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1428955 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428982 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1428982 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1428982 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429005 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429005 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429006 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429036 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429036 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429037 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429064 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429064 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429064 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429090 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429090 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429090 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429119 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429119 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429119 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429149 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429150 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429150 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429177 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429177 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429177 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429204 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429204 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429204 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429229 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429229 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429229 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429258 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429258 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429258 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429288 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429288 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429288 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429318 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429318 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429318 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429342 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429342 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429342 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429371 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429371 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429371 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429401 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429401 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429402 DEBUG (ScheduledTrigger-5344-thread-2) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429430 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 100
   [junit4]   2> 1429430 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 1429430 DEBUG (ScheduledTrigger-5344-thread-4) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 100
   [junit4]   2> 1429457 DEBUG (ScheduledTrigger-5344-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .a

[...truncated too long message...]

utput (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 891 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-solrj/test/temp/junit4-J1-20180102_144854_0539291802088035936601.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/solr-solrj/test/temp/junit4-J0-20180102_144854_05315210044699076121995.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1460 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-analysis-extras/test/temp/junit4-J0-20180102_145354_63011375217514802840654.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-analysis-extras/test/temp/junit4-J1-20180102_145354_63012471421108741380116.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 785 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-analytics/test/temp/junit4-J0-20180102_145409_126648679869208453707.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-analytics/test/temp/junit4-J1-20180102_145409_12615520415842173628268.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 553 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-clustering/test/temp/junit4-J1-20180102_145524_70213997250026180778358.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-clustering/test/temp/junit4-J0-20180102_145524_7021024804985598552442.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 1258 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-dataimporthandler-extras/test/temp/junit4-J1-20180102_145545_97717279897214712310418.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-dataimporthandler-extras/test/temp/junit4-J0-20180102_145545_9775976262655159460939.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 646 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-dataimporthandler/test/temp/junit4-J1-20180102_145556_4749925602319292102732.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-dataimporthandler/test/temp/junit4-J0-20180102_145556_47410710446400717141987.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 558 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-cell/test/temp/junit4-J1-20180102_145626_5666668828363437524445.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-cell/test/temp/junit4-J0-20180102_145626_5669734967138569568141.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 561 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-langid/test/temp/junit4-J0-20180102_145649_0439888321806760104232.syserr
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-langid/test/temp/junit4-J1-20180102_145649_04314677425128610480185.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 660 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-ltr/test/temp/junit4-J1-20180102_145700_72314721171802539250283.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-ltr/test/temp/junit4-J0-20180102_145700_72012744389793902280161.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 595 lines...]
   [junit4] JVM J1: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-uima/test/temp/junit4-J1-20180102_145735_61715439260280506097456.syserr
   [junit4] >>> JVM J1 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J1: EOF ----

[...truncated 3 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-uima/test/temp/junit4-J0-20180102_145735_61714778944920118604130.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 548 lines...]
   [junit4] JVM J0: stderr was not empty, see: /Users/jenkins/workspace/Lucene-Solr-7.x-MacOSX/solr/build/contrib/solr-velocity/test/temp/junit4-J0-20180102_145753_85917975812803591439777.syserr
   [junit4] >>> JVM J0 emitted unexpected output (verbatim) ----
   [junit4] Java HotSpot(TM) 64-Bit Server VM warning: Option UseConcMarkSweepGC was deprecated in version 9.0 and will likely be removed in a future release.
   [junit4] <<< JVM J0: EOF ----

[...truncated 32273 lines...]