You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Apache Jenkins Server <je...@builds.apache.org> on 2018/08/24 19:45:52 UTC

[JENKINS] Lucene-Solr-Tests-7.x - Build # 815 - Still Unstable

Build: https://builds.apache.org/job/Lucene-Solr-Tests-7.x/815/

1 tests failed.
FAILED:  org.apache.solr.cloud.autoscaling.sim.TestComputePlanAction.testNodeAdded

Error Message:
ComputePlanAction should have computed exactly 1 operation, but was: [org.apache.solr.client.solrj.request.CollectionAdminRequest$MoveReplica@78116e4e, org.apache.solr.client.solrj.request.CollectionAdminRequest$MoveReplica@53690551] expected:<1> but was:<2>

Stack Trace:
java.lang.AssertionError: ComputePlanAction should have computed exactly 1 operation, but was: [org.apache.solr.client.solrj.request.CollectionAdminRequest$MoveReplica@78116e4e, org.apache.solr.client.solrj.request.CollectionAdminRequest$MoveReplica@53690551] expected:<1> but was:<2>
	at __randomizedtesting.SeedInfo.seed([4DB23C39A9F7722:611875B4383CDF21]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.junit.Assert.failNotEquals(Assert.java:647)
	at org.junit.Assert.assertEquals(Assert.java:128)
	at org.junit.Assert.assertEquals(Assert.java:472)
	at org.apache.solr.cloud.autoscaling.sim.TestComputePlanAction.testNodeAdded(TestComputePlanAction.java:314)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 12664 lines...]
   [junit4] Suite: org.apache.solr.cloud.autoscaling.sim.TestComputePlanAction
   [junit4]   2> Creating dataDir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J0/temp/solr.cloud.autoscaling.sim.TestComputePlanAction_4DB23C39A9F7722-001/init-core-data-001
   [junit4]   2> 214176 WARN  (SUITE-TestComputePlanAction-seed#[4DB23C39A9F7722]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=1 numCloses=1
   [junit4]   2> 214176 INFO  (SUITE-TestComputePlanAction-seed#[4DB23C39A9F7722]-worker) [    ] o.a.s.SolrTestCaseJ4 Using PointFields (NUMERIC_POINTS_SYSPROP=true) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 214178 INFO  (SUITE-TestComputePlanAction-seed#[4DB23C39A9F7722]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (true) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)
   [junit4]   2> 214179 INFO  (SUITE-TestComputePlanAction-seed#[4DB23C39A9F7722]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> 214210 DEBUG (SUITE-TestComputePlanAction-seed#[4DB23C39A9F7722]-worker) [    ] o.a.s.c.a.s.SimClusterStateProvider --- new Overseer leader: 127.0.0.1:10000_solr
   [junit4]   2> 214223 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Adding .auto_add_replicas and .scheduled_maintenance triggers
   [junit4]   2> 214224 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 0
   [junit4]   2> 214224 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 0, lastZnodeVersion -1
   [junit4]   2> 214225 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 0
   [junit4]   2> 214228 INFO  (SUITE-TestComputePlanAction-seed#[4DB23C39A9F7722]-worker) [    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 214247 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:10000_solr]
   [junit4]   2> 214247 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testNodeLost
   [junit4]   2> 214247 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 1
   [junit4]   2> 214249 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimCloudManager === Restarting OverseerTriggerThread and clearing object cache...
   [junit4]   2> 214250 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 214251 WARN  (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Interrupted
   [junit4]   2> java.lang.InterruptedException: null
   [junit4]   2> 	at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireInterruptibly(AbstractQueuedSynchronizer.java:1220) ~[?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.locks.ReentrantLock.lockInterruptibly(ReentrantLock.java:335) ~[?:1.8.0_172]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.OverseerTriggerThread.run(OverseerTriggerThread.java:177) [java/:?]
   [junit4]   2> 	at java.lang.Thread.run(Thread.java:748) [?:1.8.0_172]
   [junit4]   2> 214252 DEBUG (ScheduledTrigger-935-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214256 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Shutting down scheduled thread pool executor now
   [junit4]   2> 214256 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Shutting down action executor now
   [junit4]   2> 214256 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Awaiting termination for action executor
   [junit4]   2> 214256 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Awaiting termination for scheduled thread pool executor
   [junit4]   2> 214256 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers ScheduledTriggers closed completely
   [junit4]   2> 214256 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread has been closed explicitly
   [junit4]   2> 214262 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Adding .auto_add_replicas and .scheduled_maintenance triggers
   [junit4]   2> 214262 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 2
   [junit4]   2> 214269 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 2, lastZnodeVersion -1
   [junit4]   2> 214269 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 2
   [junit4]   2> 214270 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:10000_solr]
   [junit4]   2> 214270 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 214271 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 2, lastZnodeVersion 2
   [junit4]   2> 214271 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214291 DEBUG (ScheduledTrigger-940-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214313 DEBUG (ScheduledTrigger-940-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214333 DEBUG (ScheduledTrigger-940-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214353 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214373 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214394 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214399 DEBUG (simCloudManagerPool-939-thread-1) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 214400 DEBUG (simCloudManagerPool-939-thread-1) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 3
   [junit4]   2> 214401 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 3
   [junit4]   2> 214402 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 214402 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 3, lastZnodeVersion 3
   [junit4]   2> 214407 DEBUG (simCloudManagerPool-939-thread-2) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 214407 DEBUG (simCloudManagerPool-939-thread-2) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 4
   [junit4]   2> 214408 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 4
   [junit4]   2> 214409 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 214409 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 4, lastZnodeVersion 4
   [junit4]   2> 214414 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214434 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214454 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214475 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214495 DEBUG (ScheduledTrigger-940-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214515 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214518 DEBUG (simCloudManagerPool-939-thread-3) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 214519 DEBUG (simCloudManagerPool-939-thread-3) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 5
   [junit4]   2> 214519 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 5
   [junit4]   2> 214520 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger node_lost_trigger - Initial livenodes: [127.0.0.1:10001_solr, 127.0.0.1:10000_solr]
   [junit4]   2> 214520 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 214520 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 5, lastZnodeVersion 5
   [junit4]   2> 214520 DEBUG (ScheduledTrigger-940-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214522 DEBUG (simCloudManagerPool-939-thread-4) [    ] o.a.s.c.o.ClusterStateMutator building a new cName: testNodeLost
   [junit4]   2> 214536 DEBUG (ScheduledTrigger-940-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214538 DEBUG (simCloudManagerPool-939-thread-4) [    ] o.a.s.c.s.c.a.PolicyHelper Creating a new session
   [junit4]   2> 214540 DEBUG (simCloudManagerPool-939-thread-4) [    ] o.a.s.c.s.c.a.PolicyHelper New session created 
   [junit4]   2> 214541 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214556 DEBUG (ScheduledTrigger-940-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214557 INFO  (simCloudManagerPool-939-thread-4) [    ] o.a.s.c.s.c.a.PolicyHelper returnSession, curr-time 2425159183 sessionWrapper.createTime 2425158339100921, this.sessionWrapper.createTime 2425158339100921 
   [junit4]   2> 214557 DEBUG (simCloudManagerPool-939-thread-4) [    ] o.a.s.c.s.c.a.PolicyHelper session set to NULL
   [junit4]   2> 214562 DEBUG (ScheduledTrigger-940-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214576 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214577 DEBUG (simCloudManagerPool-939-thread-5) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for testNodeLost / shard1
   [junit4]   2> 214582 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214592 DEBUG (simCloudManagerPool-939-thread-7) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for testNodeLost / shard1
   [junit4]   2> 214594 DEBUG (simCloudManagerPool-939-thread-7) [    ] o.a.s.c.a.s.SimClusterStateProvider -- elected new leader for testNodeLost / shard1: core_node1
   [junit4]   2> 214596 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214602 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214602 DEBUG (simCloudManagerPool-939-thread-6) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeLost / shard1
   [junit4]   2> 214603 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction -- cluster state: znodeVersion: 6
   [junit4]   2> live nodes:[127.0.0.1:10001_solr, 127.0.0.1:10000_solr]
   [junit4]   2> collections:{testNodeLost=DocCollection(testNodeLost//clusterstate.json/5)={
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"2",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "autoCreated":"true",
   [junit4]   2>   "shards":{"shard1":{
   [junit4]   2>       "replicas":{
   [junit4]   2>         "core_node1":{
   [junit4]   2>           "core":"testNodeLost_shard1_replica_n1",
   [junit4]   2>           "leader":"true",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10001_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node2":{
   [junit4]   2>           "core":"testNodeLost_shard1_replica_n2",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10000_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0}},
   [junit4]   2>       "range":"80000000-7fffffff",
   [junit4]   2>       "state":"active"}}}}
   [junit4]   2> 214605 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for testNodeLost / shard1
   [junit4]   2> 214606 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction Stopped_node : 127.0.0.1:10001_solr
   [junit4]   2> 214607 DEBUG (simCloudManagerPool-939-thread-8) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for testNodeLost / shard1
   [junit4]   2> 214607 INFO  (simCloudManagerPool-939-thread-8) [    ] o.a.s.c.ActionThrottle Throttling leader attempts - waiting for 4354ms
   [junit4]   2> 214616 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214616 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Tracking lost node: 127.0.0.1:10001_solr
   [junit4]   2> 214623 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214623 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Tracking lost node: 127.0.0.1:10001_solr
   [junit4]   2> 214637 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214643 DEBUG (ScheduledTrigger-940-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214657 DEBUG (ScheduledTrigger-940-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214663 DEBUG (ScheduledTrigger-940-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214677 DEBUG (ScheduledTrigger-940-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214685 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214694 DEBUG (simCloudManagerPool-939-thread-8) [    ] o.a.s.c.a.s.SimClusterStateProvider -- elected new leader for testNodeLost / shard1: core_node2
   [junit4]   2> 214698 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214705 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214718 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214725 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214738 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214746 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214759 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 2
   [junit4]   2> 214766 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 2
   [junit4]   2> 214766 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger firing registered processor for lost nodes: [127.0.0.1:10001_solr]
   [junit4]   2> 214766 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.ScheduledTriggers ++++++++ Cooldown inactive - processing event: {
   [junit4]   2>   "id":"89dac23546365Tdui4styjedxyl5ip4tl1xbi4z",
   [junit4]   2>   "source":"node_lost_trigger",
   [junit4]   2>   "eventTime":2425162466354021,
   [junit4]   2>   "eventType":"NODELOST",
   [junit4]   2>   "properties":{
   [junit4]   2>     "eventTimes":[2425162466354021],
   [junit4]   2>     "nodeNames":["127.0.0.1:10001_solr"]}}
   [junit4]   2> 214767 DEBUG (ScheduledTrigger-940-thread-2) [    ] o.a.s.c.a.ScheduledTriggers Pausing all triggers: [.auto_add_replicas, .scheduled_maintenance, node_lost_trigger]
   [junit4]   2> 214773 DEBUG (simCloudManagerPool-939-thread-9) [    ] o.a.s.c.o.ClusterStateMutator building a new cName: .system
   [junit4]   2> 214784 DEBUG (simCloudManagerPool-939-thread-9) [    ] o.a.s.c.s.c.a.PolicyHelper Creating a new session
   [junit4]   2> 214786 DEBUG (simCloudManagerPool-939-thread-9) [    ] o.a.s.c.s.c.a.PolicyHelper New session created 
   [junit4]   2> 214786 INFO  (simCloudManagerPool-939-thread-9) [    ] o.a.s.c.s.c.a.PolicyHelper returnSession, curr-time 2425170643 sessionWrapper.createTime 2425170605154421, this.sessionWrapper.createTime 2425170605154421 
   [junit4]   2> 214786 DEBUG (simCloudManagerPool-939-thread-9) [    ] o.a.s.c.s.c.a.PolicyHelper session set to NULL
   [junit4]   2> 214802 DEBUG (simCloudManagerPool-939-thread-10) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for .system / shard1
   [junit4]   2> 214802 DEBUG (simCloudManagerPool-939-thread-11) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for .system / shard1
   [junit4]   2> 214802 DEBUG (simCloudManagerPool-939-thread-11) [    ] o.a.s.c.a.s.SimClusterStateProvider -- elected new leader for .system / shard1: core_node1
   [junit4]   2> 214805 DEBUG (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.a.ScheduledTriggers -- processing actions for {
   [junit4]   2>   "id":"89dac23546365Tdui4styjedxyl5ip4tl1xbi4z",
   [junit4]   2>   "source":"node_lost_trigger",
   [junit4]   2>   "eventTime":2425162466354021,
   [junit4]   2>   "eventType":"NODELOST",
   [junit4]   2>   "properties":{
   [junit4]   2>     "eventTimes":[2425162466354021],
   [junit4]   2>     "_enqueue_time_":2425169704440671,
   [junit4]   2>     "nodeNames":["127.0.0.1:10001_solr"]}}
   [junit4]   2> 214805 DEBUG (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.a.ComputePlanAction -- processing event: {
   [junit4]   2>   "id":"89dac23546365Tdui4styjedxyl5ip4tl1xbi4z",
   [junit4]   2>   "source":"node_lost_trigger",
   [junit4]   2>   "eventTime":2425162466354021,
   [junit4]   2>   "eventType":"NODELOST",
   [junit4]   2>   "properties":{
   [junit4]   2>     "eventTimes":[2425162466354021],
   [junit4]   2>     "_enqueue_time_":2425169704440671,
   [junit4]   2>     "nodeNames":["127.0.0.1:10001_solr"]}} with context properties: {BEFORE_ACTION=[compute_plan]}
   [junit4]   2> 214806 DEBUG (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.s.c.a.PolicyHelper Creating a new session
   [junit4]   2> 214807 DEBUG (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.s.c.a.PolicyHelper New session created 
   [junit4]   2> 214818 INFO  (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.a.ComputePlanAction Computed Plan: action=MOVEREPLICA&collection=testNodeLost&targetNode=127.0.0.1:10002_solr&inPlaceMove=true&replica=core_node1
   [junit4]   2> 214818 INFO  (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.s.c.a.PolicyHelper returnSession, curr-time 2425172239 sessionWrapper.createTime 2425171667067571, this.sessionWrapper.createTime 2425171667067571 
   [junit4]   2> 214818 DEBUG (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.s.c.a.PolicyHelper session set to NULL
   [junit4]   2> 214820 DEBUG (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for .system / shard1
   [junit4]   2> 214823 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction -------------_ FINAL STATE --------------
   [junit4]   2> 214823 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction * Node values: {
   [junit4]   2>   "127.0.0.1:10001_solr":{
   [junit4]   2>     "sysprop.java.vendor":"Oracle Corporation",
   [junit4]   2>     "cores":1,
   [junit4]   2>     "host":"127.0.0.1",
   [junit4]   2>     "freedisk":1023,
   [junit4]   2>     "sysprop.java.version":"1.8.0_172",
   [junit4]   2>     "ip_1":"127",
   [junit4]   2>     "ip_3":"0",
   [junit4]   2>     "ip_2":"0",
   [junit4]   2>     "ip_4":"1",
   [junit4]   2>     "node":"127.0.0.1:10001_solr",
   [junit4]   2>     "metrics:solr.jvm:buffers.direct.Count":0,
   [junit4]   2>     "port":10001,
   [junit4]   2>     "heapUsage":123450000,
   [junit4]   2>     "sysLoadAvg":1.0,
   [junit4]   2>     "metrics:solr.node:ADMIN./admin/authorization.clientErrors:count":0},
   [junit4]   2>   "127.0.0.1:10000_solr":{
   [junit4]   2>     "sysprop.java.vendor":"Oracle Corporation",
   [junit4]   2>     "cores":1,
   [junit4]   2>     "host":"127.0.0.1",
   [junit4]   2>     "freedisk":999,
   [junit4]   2>     "sysprop.java.version":"1.8.0_172",
   [junit4]   2>     "ip_1":"127",
   [junit4]   2>     "ip_3":"0",
   [junit4]   2>     "ip_2":"0",
   [junit4]   2>     "ip_4":"1",
   [junit4]   2>     "node":"127.0.0.1:10000_solr",
   [junit4]   2>     "metrics:solr.jvm:buffers.direct.Count":0,
   [junit4]   2>     "port":10000,
   [junit4]   2>     "heapUsage":123450000,
   [junit4]   2>     "sysLoadAvg":1.0,
   [junit4]   2>     "metrics:solr.node:ADMIN./admin/authorization.clientErrors:count":0},
   [junit4]   2>   "127.0.0.1:10002_solr":{
   [junit4]   2>     "sysprop.java.vendor":"Oracle Corporation",
   [junit4]   2>     "cores":1,
   [junit4]   2>     "host":"127.0.0.1",
   [junit4]   2>     "freedisk":1023,
   [junit4]   2>     "sysprop.java.version":"1.8.0_172",
   [junit4]   2>     "ip_1":"127",
   [junit4]   2>     "ip_3":"0",
   [junit4]   2>     "ip_2":"0",
   [junit4]   2>     "ip_4":"1",
   [junit4]   2>     "node":"127.0.0.1:10002_solr",
   [junit4]   2>     "metrics:solr.jvm:buffers.direct.Count":0,
   [junit4]   2>     "port":10002,
   [junit4]   2>     "heapUsage":123450000,
   [junit4]   2>     "sysLoadAvg":1.0,
   [junit4]   2>     "metrics:solr.node:ADMIN./admin/authorization.clientErrors:count":0}}
   [junit4]   2> 214823 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction * Live nodes: [127.0.0.1:10000_solr]
   [junit4]   2> 214824 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction * Collection .system state: DocCollection(.system//clusterstate.json/13)={
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "autoCreated":"true",
   [junit4]   2>   "shards":{"shard1":{
   [junit4]   2>       "replicas":{"core_node1":{
   [junit4]   2>           "core":".system_shard1_replica_n1",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":2,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":11240,
   [junit4]   2>           "node_name":"127.0.0.1:10002_solr",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":1.0468065738677979E-5,
   [junit4]   2>           "SEARCHER.searcher.numDocs":2}},
   [junit4]   2>       "range":"80000000-7fffffff",
   [junit4]   2>       "state":"active"}}}
   [junit4]   2> 214824 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction * Collection testNodeLost state: DocCollection(testNodeLost//clusterstate.json/13)={
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"2",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "autoCreated":"true",
   [junit4]   2>   "shards":{"shard1":{
   [junit4]   2>       "replicas":{
   [junit4]   2>         "core_node1":{
   [junit4]   2>           "core":"testNodeLost_shard1_replica_n1",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10001_solr",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node2":{
   [junit4]   2>           "core":"testNodeLost_shard1_replica_n2",
   [junit4]   2>           "leader":"true",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10000_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0}},
   [junit4]   2>       "range":"80000000-7fffffff",
   [junit4]   2>       "state":"active"}}}
   [junit4]   2> 214824 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.SolrTestCaseJ4 ###Ending testNodeLost
   [junit4]   2> 214829 INFO  (TEST-TestComputePlanAction.testNodeLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimSolrCloudTestCase #######################################
   [junit4]   2> ############ CLUSTER STATE ############
   [junit4]   2> #######################################
   [junit4]   2> ## Live nodes:		1
   [junit4]   2> ## Empty nodes:	0
   [junit4]   2> ## Dead nodes:		2
   [junit4]   2> ##		127.0.0.1:10001_solr
   [junit4]   2> ##		127.0.0.1:10002_solr
   [junit4]   2> ## Collections:	[.system, testNodeLost]
   [junit4]   2> ## Max replicas per node:	1
   [junit4]   2> ## Min replicas per node:	1
   [junit4]   2> ## Total replicas:		1
   [junit4]   2> ## * testNodeLost		1
   [junit4]   2> ##		- active           1
   [junit4]   2> ######### Solr op counts ##########
   [junit4]   2> ##		- CREATE             1
   [junit4]   2> ##		- autoscaling        3
   [junit4]   2> ##		- update             3
   [junit4]   2> ######### Autoscaling event counts ###########
   [junit4]   2> ## * Trigger: node_lost_trigger
   [junit4]   2> ##		- BEFORE_ACTION     1
   [junit4]   2> ##		- STARTED         1
   [junit4]   2> 
   [junit4]   2> 214847 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testNodeWithMultipleReplicasLost
   [junit4]   2> 214847 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 6
   [junit4]   2> 214847 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 6
   [junit4]   2> 214847 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeLost markers
   [junit4]   2> 214848 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread   -- deleted /autoscaling/nodeLost/127.0.0.1:10002_solr
   [junit4]   2> 214848 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 214848 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 6, lastZnodeVersion 6
   [junit4]   2> 214851 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimNodeStateProvider - removing dead node values: 127.0.0.1:10001_solr
   [junit4]   2> 214851 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimNodeStateProvider - removing dead node values: 127.0.0.1:10002_solr
   [junit4]   2> 214851 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimCloudManager === Restarting OverseerTriggerThread and clearing object cache...
   [junit4]   2> 214851 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Shutting down scheduled thread pool executor now
   [junit4]   2> 214851 WARN  (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread woken up but we are closed, exiting.
   [junit4]   2> 214851 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Shutting down action executor now
   [junit4]   2> 214851 WARN  (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.a.SystemLogListener Exception sending event to collection .system
   [junit4]   2> java.io.IOException: java.lang.InterruptedException
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.sim.SimCloudManager.request(SimCloudManager.java:632) ~[test/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.SystemLogListener.onEvent(SystemLogListener.java:118) ~[java/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers$TriggerListeners.fireListeners(ScheduledTriggers.java:793) ~[java/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers$TriggerListeners.fireListeners(ScheduledTriggers.java:769) ~[java/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers.lambda$null$3(ScheduledTriggers.java:316) ~[java/:?]
   [junit4]   2> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) ~[?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_172]
   [junit4]   2> 	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209) ~[java/:?]
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_172]
   [junit4]   2> 	at java.lang.Thread.run(Thread.java:748) [?:1.8.0_172]
   [junit4]   2> Caused by: java.lang.InterruptedException
   [junit4]   2> 	at java.util.concurrent.FutureTask.awaitDone(FutureTask.java:404) ~[?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.FutureTask.get(FutureTask.java:191) ~[?:1.8.0_172]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.sim.SimCloudManager.request(SimCloudManager.java:630) ~[test/:?]
   [junit4]   2> 	... 10 more
   [junit4]   2> 214855 DEBUG (simCloudManagerPool-939-thread-13) [    ] o.a.s.c.a.s.SimClusterStateProvider -- no leader in shard1:{
   [junit4]   2>   "replicas":{"core_node1":{
   [junit4]   2>       "core":".system_shard1_replica_n1",
   [junit4]   2>       "SEARCHER.searcher.maxDoc":2,
   [junit4]   2>       "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>       "INDEX.sizeInBytes":11240,
   [junit4]   2>       "node_name":"127.0.0.1:10002_solr",
   [junit4]   2>       "state":"down",
   [junit4]   2>       "type":"NRT",
   [junit4]   2>       "INDEX.sizeInGB":1.0468065738677979E-5,
   [junit4]   2>       "SEARCHER.searcher.numDocs":2}},
   [junit4]   2>   "range":"80000000-7fffffff",
   [junit4]   2>   "state":"active"}
   [junit4]   2> 214861 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Awaiting termination for action executor
   [junit4]   2> 214862 DEBUG (simCloudManagerPool-939-thread-14) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for .system / shard1
   [junit4]   2> 214863 DEBUG (simCloudManagerPool-939-thread-15) [    ] o.a.s.c.o.ClusterStateMutator building a new cName: .system
   [junit4]   2> 214884 DEBUG (simCloudManagerPool-939-thread-16) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for .system / shard1
   [junit4]   2> 214885 DEBUG (simCloudManagerPool-939-thread-17) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for .system / shard1
   [junit4]   2> 214888 DEBUG (simCloudManagerPool-939-thread-17) [    ] o.a.s.c.a.s.SimClusterStateProvider -- elected new leader for .system / shard1: core_node2
   [junit4]   2> 214889 WARN  (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.a.ScheduledTriggers Unhandled exception executing actions
   [junit4]   2> org.apache.lucene.store.AlreadyClosedException: ScheduledTrigger node_lost_trigger has been closed.
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers$TriggerWrapper.dequeue(ScheduledTriggers.java:549) ~[java/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers.lambda$null$3(ScheduledTriggers.java:319) ~[java/:?]
   [junit4]   2> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) ~[?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_172]
   [junit4]   2> 	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209) ~[java/:?]
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_172]
   [junit4]   2> 	at java.lang.Thread.run(Thread.java:748) [?:1.8.0_172]
   [junit4]   2> 214890 DEBUG (AutoscalingActionExecutor-941-thread-1) [    ] o.a.s.c.a.ScheduledTriggers -- processing took 4228 ms for event id=89dac23546365Tdui4styjedxyl5ip4tl1xbi4z
   [junit4]   2> 214890 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Awaiting termination for scheduled thread pool executor
   [junit4]   2> 214890 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers ScheduledTriggers closed completely
   [junit4]   2> 214890 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread has been closed explicitly
   [junit4]   2> 214892 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Adding .auto_add_replicas and .scheduled_maintenance triggers
   [junit4]   2> 214892 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 7
   [junit4]   2> 214892 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 7, lastZnodeVersion -1
   [junit4]   2> 214892 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 7
   [junit4]   2> 214893 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:10000_solr]
   [junit4]   2> 214893 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 214893 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 7, lastZnodeVersion 7
   [junit4]   2> 214893 DEBUG (ScheduledTrigger-943-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214914 DEBUG (ScheduledTrigger-943-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214934 DEBUG (ScheduledTrigger-943-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214955 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214976 DEBUG (ScheduledTrigger-943-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214993 DEBUG (simCloudManagerPool-942-thread-1) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 214993 DEBUG (simCloudManagerPool-942-thread-1) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 8
   [junit4]   2> 214996 DEBUG (ScheduledTrigger-943-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 214998 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 8
   [junit4]   2> 214998 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 214998 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 8, lastZnodeVersion 8
   [junit4]   2> 215004 DEBUG (simCloudManagerPool-942-thread-2) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 215004 DEBUG (simCloudManagerPool-942-thread-2) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 9
   [junit4]   2> 215005 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 9
   [junit4]   2> 215005 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 215005 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 9, lastZnodeVersion 9
   [junit4]   2> 215016 DEBUG (ScheduledTrigger-943-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 215036 DEBUG (ScheduledTrigger-943-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 215057 DEBUG (ScheduledTrigger-943-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 215077 DEBUG (ScheduledTrigger-943-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 215099 DEBUG (ScheduledTrigger-943-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 215108 DEBUG (simCloudManagerPool-942-thread-3) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 215108 DEBUG (simCloudManagerPool-942-thread-3) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 10
   [junit4]   2> 215109 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 10
   [junit4]   2> 215109 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger node_lost_trigger - Initial livenodes: [127.0.0.1:10004_solr, 127.0.0.1:10005_solr, 127.0.0.1:10000_solr, 127.0.0.1:10003_solr]
   [junit4]   2> 215109 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 215109 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 10, lastZnodeVersion 10
   [junit4]   2> 215109 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 4
   [junit4]   2> 215110 DEBUG (simCloudManagerPool-942-thread-4) [    ] o.a.s.c.o.ClusterStateMutator building a new cName: testNodeWithMultipleReplicasLost
   [junit4]   2> 215120 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 4
   [junit4]   2> 215125 DEBUG (simCloudManagerPool-942-thread-4) [    ] o.a.s.c.s.c.a.PolicyHelper Creating a new session
   [junit4]   2> 215125 DEBUG (simCloudManagerPool-942-thread-4) [    ] o.a.s.c.s.c.a.PolicyHelper New session created 
   [junit4]   2> 215130 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 4
   [junit4]   2> 215134 INFO  (simCloudManagerPool-942-thread-4) [    ] o.a.s.c.s.c.a.PolicyHelper returnSession, curr-time 2425188027 sessionWrapper.createTime 2425187599959271, this.sessionWrapper.createTime 2425187599959271 
   [junit4]   2> 215134 DEBUG (simCloudManagerPool-942-thread-4) [    ] o.a.s.c.s.c.a.PolicyHelper session set to NULL
   [junit4]   2> 215140 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 4
   [junit4]   2> 215145 DEBUG (simCloudManagerPool-942-thread-5) [    ] o.a.s.c.a.s.SimClusterStateProvider -- no replicas in testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215145 DEBUG (simCloudManagerPool-942-thread-5) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for testNodeWithMultipleReplicasLost / shard1
   [junit4]   2> 215146 DEBUG (simCloudManagerPool-942-thread-11) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for testNodeWithMultipleReplicasLost / shard1
   [junit4]   2> 215146 DEBUG (simCloudManagerPool-942-thread-11) [    ] o.a.s.c.a.s.SimClusterStateProvider -- elected new leader for testNodeWithMultipleReplicasLost / shard1: core_node1
   [junit4]   2> 215150 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 4
   [junit4]   2> 215156 DEBUG (simCloudManagerPool-942-thread-6) [    ] o.a.s.c.a.s.SimClusterStateProvider -- no replicas in testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215156 DEBUG (simCloudManagerPool-942-thread-6) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeWithMultipleReplicasLost / shard1
   [junit4]   2> 215160 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 4
   [junit4]   2> 215167 DEBUG (simCloudManagerPool-942-thread-7) [    ] o.a.s.c.a.s.SimClusterStateProvider -- no replicas in testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215167 DEBUG (simCloudManagerPool-942-thread-7) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeWithMultipleReplicasLost / shard1
   [junit4]   2> 215170 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 4
   [junit4]   2> 215178 DEBUG (simCloudManagerPool-942-thread-8) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215179 DEBUG (simCloudManagerPool-942-thread-8) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeWithMultipleReplicasLost / shard1
   [junit4]   2> 215179 DEBUG (simCloudManagerPool-942-thread-12) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215179 DEBUG (simCloudManagerPool-942-thread-12) [    ] o.a.s.c.a.s.SimClusterStateProvider -- elected new leader for testNodeWithMultipleReplicasLost / shard2: core_node4
   [junit4]   2> 215180 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 4
   [junit4]   2> 215189 DEBUG (simCloudManagerPool-942-thread-9) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215189 DEBUG (simCloudManagerPool-942-thread-9) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeWithMultipleReplicasLost / shard1
   [junit4]   2> 215190 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 4
   [junit4]   2> 215200 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 4
   [junit4]   2> 215202 DEBUG (simCloudManagerPool-942-thread-10) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215202 DEBUG (simCloudManagerPool-942-thread-10) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeWithMultipleReplicasLost / shard1
   [junit4]   2> 215202 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction -- cluster state: znodeVersion: 28
   [junit4]   2> live nodes:[127.0.0.1:10004_solr, 127.0.0.1:10000_solr, 127.0.0.1:10003_solr, 127.0.0.1:10005_solr]
   [junit4]   2> collections:{.system=DocCollection(.system//clusterstate.json/27)={
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "autoCreated":"true",
   [junit4]   2>   "shards":{"shard1":{
   [junit4]   2>       "replicas":{"core_node2":{
   [junit4]   2>           "core":".system_shard1_replica_n1",
   [junit4]   2>           "leader":"true",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":1,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10740,
   [junit4]   2>           "node_name":"127.0.0.1:10000_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":1.000240445137024E-5,
   [junit4]   2>           "SEARCHER.searcher.numDocs":1}},
   [junit4]   2>       "range":"80000000-7fffffff",
   [junit4]   2>       "state":"active"}}}, testNodeWithMultipleReplicasLost=DocCollection(testNodeWithMultipleReplicasLost//clusterstate.json/27)={
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"3",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "autoCreated":"true",
   [junit4]   2>   "shards":{
   [junit4]   2>     "shard2":{
   [junit4]   2>       "replicas":{
   [junit4]   2>         "core_node6":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard2_replica_n6",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10003_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node4":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard2_replica_n4",
   [junit4]   2>           "leader":"true",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10004_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node5":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard2_replica_n5",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10005_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0}},
   [junit4]   2>       "range":"0-7fffffff",
   [junit4]   2>       "state":"active"},
   [junit4]   2>     "shard1":{
   [junit4]   2>       "replicas":{
   [junit4]   2>         "core_node1":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard1_replica_n1",
   [junit4]   2>           "leader":"true",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10003_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node2":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard1_replica_n2",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10005_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node3":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard1_replica_n3",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10004_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0}},
   [junit4]   2>       "range":"80000000-ffffffff",
   [junit4]   2>       "state":"active"}}}}
   [junit4]   2> 215204 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215205 DEBUG (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimClusterStateProvider -- already has leader testNodeWithMultipleReplicasLost / shard1
   [junit4]   2> 215205 DEBUG (simCloudManagerPool-942-thread-13) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for testNodeWithMultipleReplicasLost / shard2
   [junit4]   2> 215205 INFO  (simCloudManagerPool-942-thread-13) [    ] o.a.s.c.ActionThrottle Throttling leader attempts - waiting for 3701ms
   [junit4]   2> 215211 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 3
   [junit4]   2> 215211 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Tracking lost node: 127.0.0.1:10004_solr
   [junit4]   2> 215221 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 3
   [junit4]   2> 215221 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Tracking lost node: 127.0.0.1:10004_solr
   [junit4]   2> 215232 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: node_lost_trigger with currently live nodes: 3
   [junit4]   2> 215232 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger firing registered processor for lost nodes: [127.0.0.1:10004_solr]
   [junit4]   2> 215232 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.ScheduledTriggers ++++++++ Cooldown inactive - processing event: {
   [junit4]   2>   "id":"89db2fd300403Tdui4styjedxyl5ip4tl1xbi54",
   [junit4]   2>   "source":"node_lost_trigger",
   [junit4]   2>   "eventTime":2425191891207171,
   [junit4]   2>   "eventType":"NODELOST",
   [junit4]   2>   "properties":{
   [junit4]   2>     "eventTimes":[2425191891207171],
   [junit4]   2>     "nodeNames":["127.0.0.1:10004_solr"]}}
   [junit4]   2> 215232 DEBUG (ScheduledTrigger-943-thread-3) [    ] o.a.s.c.a.ScheduledTriggers Pausing all triggers: [.auto_add_replicas, .scheduled_maintenance, node_lost_trigger]
   [junit4]   2> 215235 DEBUG (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.a.ScheduledTriggers -- processing actions for {
   [junit4]   2>   "id":"89db2fd300403Tdui4styjedxyl5ip4tl1xbi54",
   [junit4]   2>   "source":"node_lost_trigger",
   [junit4]   2>   "eventTime":2425191891207171,
   [junit4]   2>   "eventType":"NODELOST",
   [junit4]   2>   "properties":{
   [junit4]   2>     "eventTimes":[2425191891207171],
   [junit4]   2>     "_enqueue_time_":2425192956901871,
   [junit4]   2>     "nodeNames":["127.0.0.1:10004_solr"]}}
   [junit4]   2> 215235 DEBUG (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.a.ComputePlanAction -- processing event: {
   [junit4]   2>   "id":"89db2fd300403Tdui4styjedxyl5ip4tl1xbi54",
   [junit4]   2>   "source":"node_lost_trigger",
   [junit4]   2>   "eventTime":2425191891207171,
   [junit4]   2>   "eventType":"NODELOST",
   [junit4]   2>   "properties":{
   [junit4]   2>     "eventTimes":[2425191891207171],
   [junit4]   2>     "_enqueue_time_":2425192956901871,
   [junit4]   2>     "nodeNames":["127.0.0.1:10004_solr"]}} with context properties: {BEFORE_ACTION=[compute_plan]}
   [junit4]   2> 215236 DEBUG (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.s.c.a.PolicyHelper Creating a new session
   [junit4]   2> 215239 DEBUG (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.s.c.a.PolicyHelper New session created 
   [junit4]   2> 215255 INFO  (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.a.ComputePlanAction Computed Plan: action=MOVEREPLICA&collection=testNodeWithMultipleReplicasLost&targetNode=127.0.0.1:10000_solr&inPlaceMove=true&replica=core_node4
   [junit4]   2> 215256 INFO  (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.a.ComputePlanAction Computed Plan: action=MOVEREPLICA&collection=testNodeWithMultipleReplicasLost&targetNode=127.0.0.1:10000_solr&inPlaceMove=true&replica=core_node3
   [junit4]   2> 215256 INFO  (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.s.c.a.PolicyHelper returnSession, curr-time 2425194134 sessionWrapper.createTime 2425193286357121, this.sessionWrapper.createTime 2425193286357121 
   [junit4]   2> 215256 DEBUG (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.s.c.a.PolicyHelper session set to NULL
   [junit4]   2> 215274 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction action=MOVEREPLICA&collection=testNodeWithMultipleReplicasLost&targetNode=127.0.0.1:10000_solr&inPlaceMove=true&replica=core_node4
   [junit4]   2> 215274 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction action=MOVEREPLICA&collection=testNodeWithMultipleReplicasLost&targetNode=127.0.0.1:10000_solr&inPlaceMove=true&replica=core_node3
   [junit4]   2> 215274 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction -------------_ FINAL STATE --------------
   [junit4]   2> 215275 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction * Node values: {
   [junit4]   2>   "127.0.0.1:10004_solr":{
   [junit4]   2>     "sysprop.java.vendor":"Oracle Corporation",
   [junit4]   2>     "cores":2,
   [junit4]   2>     "host":"127.0.0.1",
   [junit4]   2>     "freedisk":1022,
   [junit4]   2>     "sysprop.java.version":"1.8.0_172",
   [junit4]   2>     "ip_1":"127",
   [junit4]   2>     "ip_3":"0",
   [junit4]   2>     "ip_2":"0",
   [junit4]   2>     "ip_4":"1",
   [junit4]   2>     "node":"127.0.0.1:10004_solr",
   [junit4]   2>     "metrics:solr.jvm:buffers.direct.Count":0,
   [junit4]   2>     "port":10004,
   [junit4]   2>     "heapUsage":123450000,
   [junit4]   2>     "sysLoadAvg":1.0,
   [junit4]   2>     "metrics:solr.node:ADMIN./admin/authorization.clientErrors:count":0},
   [junit4]   2>   "127.0.0.1:10005_solr":{
   [junit4]   2>     "sysprop.java.vendor":"Oracle Corporation",
   [junit4]   2>     "cores":2,
   [junit4]   2>     "host":"127.0.0.1",
   [junit4]   2>     "freedisk":1022,
   [junit4]   2>     "sysprop.java.version":"1.8.0_172",
   [junit4]   2>     "ip_1":"127",
   [junit4]   2>     "ip_3":"0",
   [junit4]   2>     "ip_2":"0",
   [junit4]   2>     "ip_4":"1",
   [junit4]   2>     "node":"127.0.0.1:10005_solr",
   [junit4]   2>     "metrics:solr.jvm:buffers.direct.Count":0,
   [junit4]   2>     "port":10005,
   [junit4]   2>     "heapUsage":123450000,
   [junit4]   2>     "sysLoadAvg":1.0,
   [junit4]   2>     "metrics:solr.node:ADMIN./admin/authorization.clientErrors:count":0},
   [junit4]   2>   "127.0.0.1:10000_solr":{
   [junit4]   2>     "sysprop.java.vendor":"Oracle Corporation",
   [junit4]   2>     "cores":1,
   [junit4]   2>     "host":"127.0.0.1",
   [junit4]   2>     "freedisk":999,
   [junit4]   2>     "sysprop.java.version":"1.8.0_172",
   [junit4]   2>     "ip_1":"127",
   [junit4]   2>     "ip_3":"0",
   [junit4]   2>     "ip_2":"0",
   [junit4]   2>     "ip_4":"1",
   [junit4]   2>     "node":"127.0.0.1:10000_solr",
   [junit4]   2>     "metrics:solr.jvm:buffers.direct.Count":0,
   [junit4]   2>     "port":10000,
   [junit4]   2>     "heapUsage":123450000,
   [junit4]   2>     "sysLoadAvg":1.0,
   [junit4]   2>     "metrics:solr.node:ADMIN./admin/authorization.clientErrors:count":0},
   [junit4]   2>   "127.0.0.1:10003_solr":{
   [junit4]   2>     "sysprop.java.vendor":"Oracle Corporation",
   [junit4]   2>     "cores":2,
   [junit4]   2>     "host":"127.0.0.1",
   [junit4]   2>     "freedisk":1022,
   [junit4]   2>     "sysprop.java.version":"1.8.0_172",
   [junit4]   2>     "ip_1":"127",
   [junit4]   2>     "ip_3":"0",
   [junit4]   2>     "ip_2":"0",
   [junit4]   2>     "ip_4":"1",
   [junit4]   2>     "node":"127.0.0.1:10003_solr",
   [junit4]   2>     "metrics:solr.jvm:buffers.direct.Count":0,
   [junit4]   2>     "port":10003,
   [junit4]   2>     "heapUsage":123450000,
   [junit4]   2>     "sysLoadAvg":1.0,
   [junit4]   2>     "metrics:solr.node:ADMIN./admin/authorization.clientErrors:count":0}}
   [junit4]   2> 215275 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction * Live nodes: [127.0.0.1:10005_solr, 127.0.0.1:10000_solr, 127.0.0.1:10003_solr]
   [junit4]   2> 215275 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction * Collection .system state: DocCollection(.system//clusterstate.json/32)={
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "autoCreated":"true",
   [junit4]   2>   "shards":{"shard1":{
   [junit4]   2>       "replicas":{"core_node2":{
   [junit4]   2>           "core":".system_shard1_replica_n1",
   [junit4]   2>           "leader":"true",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":3,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":11740,
   [junit4]   2>           "node_name":"127.0.0.1:10000_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":1.0933727025985718E-5,
   [junit4]   2>           "SEARCHER.searcher.numDocs":3}},
   [junit4]   2>       "range":"80000000-7fffffff",
   [junit4]   2>       "state":"active"}}}
   [junit4]   2> 215276 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.TestComputePlanAction * Collection testNodeWithMultipleReplicasLost state: DocCollection(testNodeWithMultipleReplicasLost//clusterstate.json/32)={
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"3",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "autoCreated":"true",
   [junit4]   2>   "shards":{
   [junit4]   2>     "shard2":{
   [junit4]   2>       "replicas":{
   [junit4]   2>         "core_node6":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard2_replica_n6",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10003_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node4":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard2_replica_n4",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10004_solr",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node5":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard2_replica_n5",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10005_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0}},
   [junit4]   2>       "range":"0-7fffffff",
   [junit4]   2>       "state":"active"},
   [junit4]   2>     "shard1":{
   [junit4]   2>       "replicas":{
   [junit4]   2>         "core_node1":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard1_replica_n1",
   [junit4]   2>           "leader":"true",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10003_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node2":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard1_replica_n2",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10005_solr",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0},
   [junit4]   2>         "core_node3":{
   [junit4]   2>           "core":"testNodeWithMultipleReplicasLost_shard1_replica_n3",
   [junit4]   2>           "SEARCHER.searcher.maxDoc":0,
   [junit4]   2>           "SEARCHER.searcher.deletedDocs":0,
   [junit4]   2>           "INDEX.sizeInBytes":10240,
   [junit4]   2>           "node_name":"127.0.0.1:10004_solr",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "INDEX.sizeInGB":9.5367431640625E-6,
   [junit4]   2>           "SEARCHER.searcher.numDocs":0}},
   [junit4]   2>       "range":"80000000-ffffffff",
   [junit4]   2>       "state":"active"}}}
   [junit4]   2> 215276 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.SolrTestCaseJ4 ###Ending testNodeWithMultipleReplicasLost
   [junit4]   2> 215276 INFO  (TEST-TestComputePlanAction.testNodeWithMultipleReplicasLost-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimSolrCloudTestCase #######################################
   [junit4]   2> ############ CLUSTER STATE ############
   [junit4]   2> #######################################
   [junit4]   2> ## Live nodes:		3
   [junit4]   2> ## Empty nodes:	0
   [junit4]   2> ## Dead nodes:		1
   [junit4]   2> ##		127.0.0.1:10004_solr
   [junit4]   2> ## Collections:	[.system, testNodeWithMultipleReplicasLost]
   [junit4]   2> ## Max replicas per node:	2
   [junit4]   2> ## Min replicas per node:	1
   [junit4]   2> ## Total replicas:		5
   [junit4]   2> ## * .system		1
   [junit4]   2> ##		- active           1
   [junit4]   2> ## * testNodeWithMultipleReplicasLost		4
   [junit4]   2> ##		- active           4
   [junit4]   2> ######### Solr op counts ##########
   [junit4]   2> ##		- CREATE             1
   [junit4]   2> ##		- autoscaling        3
   [junit4]   2> ##		- update             2
   [junit4]   2> ######### Autoscaling event counts ###########
   [junit4]   2> ## * Trigger: node_lost_trigger
   [junit4]   2> ##		- AFTER_ACTION     1
   [junit4]   2> ##		- BEFORE_ACTION     1
   [junit4]   2> ##		- FAILED          1
   [junit4]   2> ##		- STARTED         1
   [junit4]   2> 
   [junit4]   2> 215280 DEBUG (simCloudManagerPool-942-thread-13) [    ] o.a.s.c.a.s.SimClusterStateProvider -- elected new leader for testNodeWithMultipleReplicasLost / shard2: core_node6
   [junit4]   2> 215291 INFO  (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testNodeAdded
   [junit4]   2> 215291 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 11
   [junit4]   2> 215291 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimNodeStateProvider - removing dead node values: 127.0.0.1:10004_solr
   [junit4]   2> 215291 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimNodeStateProvider --removing value for 127.0.0.1:10005_solr
   [junit4]   2> 215291 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimNodeStateProvider --removing value for 127.0.0.1:10003_solr
   [junit4]   2> 215291 INFO  (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.s.SimCloudManager === Restarting OverseerTriggerThread and clearing object cache...
   [junit4]   2> 215291 WARN  (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread woken up but we are closed, exiting.
   [junit4]   2> 215292 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Shutting down scheduled thread pool executor now
   [junit4]   2> 215292 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Shutting down action executor now
   [junit4]   2> 215292 DEBUG (simCloudManagerPool-942-thread-16) [    ] o.a.s.c.o.ClusterStateMutator building a new cName: .system
   [junit4]   2> 215292 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Awaiting termination for action executor
   [junit4]   2> 215293 WARN  (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.a.SystemLogListener Exception sending event to collection .system
   [junit4]   2> java.io.IOException: java.lang.InterruptedException
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.sim.SimCloudManager.request(SimCloudManager.java:632) ~[test/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.SystemLogListener.onEvent(SystemLogListener.java:118) ~[java/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers$TriggerListeners.fireListeners(ScheduledTriggers.java:793) ~[java/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers$TriggerListeners.fireListeners(ScheduledTriggers.java:769) ~[java/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers.lambda$null$3(ScheduledTriggers.java:316) ~[java/:?]
   [junit4]   2> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) ~[?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_172]
   [junit4]   2> 	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209) ~[java/:?]
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_172]
   [junit4]   2> 	at java.lang.Thread.run(Thread.java:748) [?:1.8.0_172]
   [junit4]   2> Caused by: java.lang.InterruptedException
   [junit4]   2> 	at java.util.concurrent.FutureTask.awaitDone(FutureTask.java:404) ~[?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.FutureTask.get(FutureTask.java:191) ~[?:1.8.0_172]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.sim.SimCloudManager.request(SimCloudManager.java:630) ~[test/:?]
   [junit4]   2> 	... 10 more
   [junit4]   2> 215294 DEBUG (simCloudManagerPool-942-thread-17) [    ] o.a.s.c.o.ClusterStateMutator building a new cName: .system
   [junit4]   2> 215314 DEBUG (simCloudManagerPool-942-thread-18) [    ] o.a.s.c.a.s.SimClusterStateProvider -- submit leader election for .system / shard1
   [junit4]   2> 215321 DEBUG (simCloudManagerPool-942-thread-19) [    ] o.a.s.c.a.s.SimClusterStateProvider Running leader election for .system / shard1
   [junit4]   2> 215321 DEBUG (simCloudManagerPool-942-thread-19) [    ] o.a.s.c.a.s.SimClusterStateProvider -- elected new leader for .system / shard1: core_node3
   [junit4]   2> 216522 WARN  (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.a.ScheduledTriggers Unhandled exception executing actions
   [junit4]   2> org.apache.lucene.store.AlreadyClosedException: ScheduledTrigger node_lost_trigger has been closed.
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers$TriggerWrapper.dequeue(ScheduledTriggers.java:549) ~[java/:?]
   [junit4]   2> 	at org.apache.solr.cloud.autoscaling.ScheduledTriggers.lambda$null$3(ScheduledTriggers.java:319) ~[java/:?]
   [junit4]   2> 	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) ~[?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_172]
   [junit4]   2> 	at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209) ~[java/:?]
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_172]
   [junit4]   2> 	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_172]
   [junit4]   2> 	at java.lang.Thread.run(Thread.java:748) [?:1.8.0_172]
   [junit4]   2> 216523 DEBUG (AutoscalingActionExecutor-944-thread-1) [    ] o.a.s.c.a.ScheduledTriggers -- processing took 64367 ms for event id=89db2fd300403Tdui4styjedxyl5ip4tl1xbi54
   [junit4]   2> 216523 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers Awaiting termination for scheduled thread pool executor
   [junit4]   2> 216523 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.ScheduledTriggers ScheduledTriggers closed completely
   [junit4]   2> 216523 DEBUG (TEST-TestComputePlanAction.testNodeAdded-seed#[4DB23C39A9F7722]) [    ] o.a.s.c.a.OverseerTriggerThread OverseerTriggerThread has been closed explicitly
   [junit4]   2> 216525 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Adding .auto_add_replicas and .scheduled_maintenance triggers
   [junit4]   2> 216525 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 12
   [junit4]   2> 216525 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 12, lastZnodeVersion -1
   [junit4]   2> 216525 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 12
   [junit4]   2> 216528 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeLostTrigger NodeLostTrigger .auto_add_replicas - Initial livenodes: [127.0.0.1:10000_solr]
   [junit4]   2> 216528 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 216528 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 12, lastZnodeVersion 12
   [junit4]   2> 216528 DEBUG (ScheduledTrigger-946-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216549 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216569 DEBUG (ScheduledTrigger-946-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216589 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216609 DEBUG (ScheduledTrigger-946-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216628 DEBUG (simCloudManagerPool-945-thread-1) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 216629 DEBUG (simCloudManagerPool-945-thread-1) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 13
   [junit4]   2> 216630 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216646 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 13
   [junit4]   2> 216646 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 216646 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 13, lastZnodeVersion 13
   [junit4]   2> 216650 DEBUG (ScheduledTrigger-946-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216654 DEBUG (simCloudManagerPool-945-thread-6) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 216654 DEBUG (simCloudManagerPool-945-thread-6) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 14
   [junit4]   2> 216654 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 14
   [junit4]   2> 216654 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread -- clean old nodeAdded markers
   [junit4]   2> 216655 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 14, lastZnodeVersion 14
   [junit4]   2> 216670 DEBUG (ScheduledTrigger-946-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216691 DEBUG (ScheduledTrigger-946-thread-4) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216711 DEBUG (ScheduledTrigger-946-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216731 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216751 DEBUG (ScheduledTrigger-946-thread-3) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216771 DEBUG (ScheduledTrigger-946-thread-2) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216785 DEBUG (simCloudManagerPool-945-thread-7) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 216785 DEBUG (simCloudManagerPool-945-thread-7) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 15
   [junit4]   2> 216786 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 15
   [junit4]   2> 216786 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeAddedTrigger NodeAddedTrigger node_added_trigger - Initial livenodes: [127.0.0.1:10000_solr]
   [junit4]   2> 216786 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.NodeAddedTrigger NodeAddedTrigger node_added_trigger instantiated with properties: {event=nodeAdded, waitFor=1, actions=[{name=compute_plan, class=solr.ComputePlanAction}, {name=test, class=org.apache.solr.cloud.autoscaling.sim.TestComputePlanAction$AssertingTriggerAction}], enabled=true}
   [junit4]   2> 216787 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Current znodeVersion 15, lastZnodeVersion 15
   [junit4]   2> 216787 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 216787 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 1
   [junit4]   2> 216792 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216807 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Running NodeAddedTrigger node_added_trigger
   [junit4]   2> 216807 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeAddedTrigger Found livenodes: 1
   [junit4]   2> 216812 DEBUG (ScheduledTrigger-946-thread-1) [    ] o.a.s.c.a.NodeLostTrigger Running NodeLostTrigger: .auto_add_replicas with currently live nodes: 1
   [junit4]   2> 216817 DEBUG (simCloudManagerPool-945-thread-8) [    ] o.a.s.c.a.AutoScalingHandler Verified autoscaling configuration
   [junit4]   2> 216817 DEBUG (simCloudManagerPool-945-thread-8) [    ] o.a.s.c.a.OverseerTriggerThread Refreshing /autoscaling.json with znode version 16
   [junit4]   2> 216817 DEBUG (Simulated OverseerAutoScalingTriggerThread) [    ] o.a.s.c.a.OverseerTriggerThread Processed trigger updates upto znodeVersion 16
   [junit4]   2> 216818 DEBUG (Simulated Ov

[...truncated too long message...]

eck:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

jar-checksums:
    [mkdir] Created dir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1505158017
     [copy] Copying 239 files to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1505158017
   [delete] Deleting directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1505158017

check-working-copy:
[ivy:cachepath] :: resolving dependencies :: org.eclipse.jgit#org.eclipse.jgit-caller;working
[ivy:cachepath] 	confs: [default]
[ivy:cachepath] 	found org.eclipse.jgit#org.eclipse.jgit;4.6.0.201612231935-r in public
[ivy:cachepath] 	found com.jcraft#jsch;0.1.53 in public
[ivy:cachepath] 	found com.googlecode.javaewah#JavaEWAH;1.1.6 in public
[ivy:cachepath] 	found org.apache.httpcomponents#httpclient;4.3.6 in public
[ivy:cachepath] 	found org.apache.httpcomponents#httpcore;4.3.3 in public
[ivy:cachepath] 	found commons-logging#commons-logging;1.1.3 in public
[ivy:cachepath] 	found commons-codec#commons-codec;1.6 in public
[ivy:cachepath] 	found org.slf4j#slf4j-api;1.7.2 in public
[ivy:cachepath] :: resolution report :: resolve 25ms :: artifacts dl 1ms
	---------------------------------------------------------------------
	|                  |            modules            ||   artifacts   |
	|       conf       | number| search|dwnlded|evicted|| number|dwnlded|
	---------------------------------------------------------------------
	|      default     |   8   |   0   |   0   |   0   ||   8   |   0   |
	---------------------------------------------------------------------
[wc-checker] Initializing working copy...
[wc-checker] SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
[wc-checker] SLF4J: Defaulting to no-operation (NOP) logger implementation
[wc-checker] SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.
[wc-checker] Checking working copy status...

-jenkins-base:

BUILD SUCCESSFUL
Total time: 91 minutes 22 seconds
Archiving artifacts
java.lang.InterruptedException: no matches found within 10000
	at hudson.FilePath$34.hasMatch(FilePath.java:2678)
	at hudson.FilePath$34.invoke(FilePath.java:2557)
	at hudson.FilePath$34.invoke(FilePath.java:2547)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2918)
Also:   hudson.remoting.Channel$CallSiteStackTrace: Remote call to lucene
		at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
		at hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
		at hudson.remoting.Channel.call(Channel.java:955)
		at hudson.FilePath.act(FilePath.java:1036)
		at hudson.FilePath.act(FilePath.java:1025)
		at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
		at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
		at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
		at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
		at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
		at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
		at hudson.model.Build$BuildExecution.post2(Build.java:186)
		at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
		at hudson.model.Run.execute(Run.java:1819)
		at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
		at hudson.model.ResourceController.execute(ResourceController.java:97)
		at hudson.model.Executor.run(Executor.java:429)
Caused: hudson.FilePath$TunneledInterruptedException
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2920)
	at hudson.remoting.UserRequest.perform(UserRequest.java:212)
	at hudson.remoting.UserRequest.perform(UserRequest.java:54)
	at hudson.remoting.Request$2.run(Request.java:369)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:748)
Caused: java.lang.InterruptedException: java.lang.InterruptedException: no matches found within 10000
	at hudson.FilePath.act(FilePath.java:1038)
	at hudson.FilePath.act(FilePath.java:1025)
	at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
	at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
	at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
	at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
	at hudson.model.Build$BuildExecution.post2(Build.java:186)
	at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
	at hudson.model.Run.execute(Run.java:1819)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
No artifacts found that match the file pattern "**/*.events,heapdumps/**,**/hs_err_pid*". Configuration error?
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)

[JENKINS] Lucene-Solr-Tests-7.x - Build # 817 - Still Unstable

Posted by Apache Jenkins Server <je...@builds.apache.org>.
Build: https://builds.apache.org/job/Lucene-Solr-Tests-7.x/817/

2 tests failed.
FAILED:  org.apache.solr.cloud.MoveReplicaHDFSTest.testFailedMove

Error Message:
No live SolrServers available to handle this request:[http://127.0.0.1:36278/solr/MoveReplicaHDFSTest_failed_coll_true, http://127.0.0.1:36493/solr/MoveReplicaHDFSTest_failed_coll_true]

Stack Trace:
org.apache.solr.client.solrj.SolrServerException: No live SolrServers available to handle this request:[http://127.0.0.1:36278/solr/MoveReplicaHDFSTest_failed_coll_true, http://127.0.0.1:36493/solr/MoveReplicaHDFSTest_failed_coll_true]
	at __randomizedtesting.SeedInfo.seed([DDC8408B1482A6D5:77059379A3517305]:0)
	at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:462)
	at org.apache.solr.client.solrj.impl.CloudSolrClient.sendRequest(CloudSolrClient.java:1109)
	at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:886)
	at org.apache.solr.client.solrj.impl.CloudSolrClient.requestWithRetryOnStaleState(CloudSolrClient.java:996)
	at org.apache.solr.client.solrj.impl.CloudSolrClient.request(CloudSolrClient.java:819)
	at org.apache.solr.client.solrj.SolrRequest.process(SolrRequest.java:194)
	at org.apache.solr.client.solrj.SolrClient.query(SolrClient.java:942)
	at org.apache.solr.cloud.MoveReplicaTest.testFailedMove(MoveReplicaTest.java:289)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at http://127.0.0.1:36278/solr/MoveReplicaHDFSTest_failed_coll_true: no servers hosting shard: shard1
	at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:643)
	at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255)
	at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244)
	at org.apache.solr.client.solrj.impl.LBHttpSolrClient.doRequest(LBHttpSolrClient.java:483)
	at org.apache.solr.client.solrj.impl.LBHttpSolrClient.request(LBHttpSolrClient.java:436)
	... 46 more


FAILED:  org.apache.solr.cloud.autoscaling.IndexSizeTriggerTest.testSplitIntegration

Error Message:


Stack Trace:
java.util.ConcurrentModificationException
	at __randomizedtesting.SeedInfo.seed([DDC8408B1482A6D5:E446F9CB3B7D6F2B]:0)
	at java.util.ArrayList$Itr.checkForComodification(ArrayList.java:909)
	at java.util.ArrayList$Itr.next(ArrayList.java:859)
	at java.util.AbstractCollection.toString(AbstractCollection.java:461)
	at java.lang.String.valueOf(String.java:2994)
	at java.lang.StringBuilder.append(StringBuilder.java:131)
	at org.apache.solr.cloud.autoscaling.IndexSizeTriggerTest.testSplitIntegration(IndexSizeTriggerTest.java:310)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 13359 lines...]
   [junit4] Suite: org.apache.solr.cloud.MoveReplicaHDFSTest
   [junit4]   2> 811365 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.SolrTestCaseJ4 SecureRandom sanity checks: test.solr.allowed.securerandom=null & java.security.egd=file:/dev/./urandom
   [junit4]   2> Creating dataDir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/init-core-data-001
   [junit4]   2> 811366 WARN  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.SolrTestCaseJ4 startTrackingSearchers: numOpens=1 numCloses=1
   [junit4]   2> 811366 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.SolrTestCaseJ4 Using TrieFields (NUMERIC_POINTS_SYSPROP=false) w/NUMERIC_DOCVALUES_SYSPROP=true
   [junit4]   2> 811367 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.SolrTestCaseJ4 Randomized ssl (false) and clientAuth (false) via: @org.apache.solr.util.RandomizeSSL(reason=, ssl=NaN, value=NaN, clientAuth=NaN)
   [junit4]   2> 811368 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.c.MiniSolrCloudCluster Starting cluster of 4 servers in /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001
   [junit4]   2> 811368 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 811374 INFO  (Thread-2335) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 811374 INFO  (Thread-2335) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 811376 ERROR (Thread-2335) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 811474 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.c.ZkTestServer start zk server on port:44880
   [junit4]   2> 811476 INFO  (zkConnectionManagerCallback-2767-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811479 INFO  (jetty-launcher-2764-thread-1) [    ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
   [junit4]   2> 811480 INFO  (jetty-launcher-2764-thread-2) [    ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
   [junit4]   2> 811481 INFO  (jetty-launcher-2764-thread-3) [    ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
   [junit4]   2> 811481 INFO  (jetty-launcher-2764-thread-4) [    ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
   [junit4]   2> 811481 INFO  (jetty-launcher-2764-thread-1) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 811481 INFO  (jetty-launcher-2764-thread-1) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 811481 INFO  (jetty-launcher-2764-thread-1) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 811482 INFO  (jetty-launcher-2764-thread-3) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 811482 INFO  (jetty-launcher-2764-thread-3) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 811482 INFO  (jetty-launcher-2764-thread-3) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 811482 INFO  (jetty-launcher-2764-thread-3) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@f949a56{/solr,null,AVAILABLE}
   [junit4]   2> 811483 INFO  (jetty-launcher-2764-thread-2) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 811483 INFO  (jetty-launcher-2764-thread-2) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 811483 INFO  (jetty-launcher-2764-thread-2) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 811483 INFO  (jetty-launcher-2764-thread-1) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@34eeb45b{/solr,null,AVAILABLE}
   [junit4]   2> 811483 INFO  (jetty-launcher-2764-thread-2) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@245f01b4{/solr,null,AVAILABLE}
   [junit4]   2> 811483 INFO  (jetty-launcher-2764-thread-1) [    ] o.e.j.s.AbstractConnector Started ServerConnector@395cf0a1{HTTP/1.1,[http/1.1]}{127.0.0.1:34594}
   [junit4]   2> 811484 INFO  (jetty-launcher-2764-thread-1) [    ] o.e.j.s.Server Started @811572ms
   [junit4]   2> 811484 INFO  (jetty-launcher-2764-thread-1) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=34594}
   [junit4]   2> 811484 ERROR (jetty-launcher-2764-thread-1) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 811484 INFO  (jetty-launcher-2764-thread-1) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 811484 INFO  (jetty-launcher-2764-thread-1) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solrâ„¢ version 7.5.0
   [junit4]   2> 811484 INFO  (jetty-launcher-2764-thread-1) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 811484 INFO  (jetty-launcher-2764-thread-1) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 811484 INFO  (jetty-launcher-2764-thread-1) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-08-25T13:21:25.279Z
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-3) [    ] o.e.j.s.AbstractConnector Started ServerConnector@12b568f8{HTTP/1.1,[http/1.1]}{127.0.0.1:36278}
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-3) [    ] o.e.j.s.Server Started @811573ms
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-3) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=36278}
   [junit4]   2> 811485 ERROR (jetty-launcher-2764-thread-3) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-3) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-3) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solrâ„¢ version 7.5.0
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-3) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-3) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-3) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-08-25T13:21:25.280Z
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-4) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 811485 INFO  (jetty-launcher-2764-thread-4) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@42ed1b48{/solr,null,AVAILABLE}
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.e.j.s.AbstractConnector Started ServerConnector@5cc1603c{HTTP/1.1,[http/1.1]}{127.0.0.1:44893}
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.e.j.s.Server Started @811575ms
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=44893}
   [junit4]   2> 811486 ERROR (jetty-launcher-2764-thread-4) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solrâ„¢ version 7.5.0
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 811486 INFO  (jetty-launcher-2764-thread-4) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-08-25T13:21:25.281Z
   [junit4]   2> 811488 INFO  (jetty-launcher-2764-thread-2) [    ] o.e.j.s.AbstractConnector Started ServerConnector@5a22441e{HTTP/1.1,[http/1.1]}{127.0.0.1:42205}
   [junit4]   2> 811488 INFO  (jetty-launcher-2764-thread-2) [    ] o.e.j.s.Server Started @811576ms
   [junit4]   2> 811488 INFO  (jetty-launcher-2764-thread-2) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=42205}
   [junit4]   2> 811488 ERROR (jetty-launcher-2764-thread-2) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 811488 INFO  (jetty-launcher-2764-thread-2) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 811488 INFO  (jetty-launcher-2764-thread-2) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solrâ„¢ version 7.5.0
   [junit4]   2> 811488 INFO  (jetty-launcher-2764-thread-2) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 811488 INFO  (jetty-launcher-2764-thread-2) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 811488 INFO  (jetty-launcher-2764-thread-2) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-08-25T13:21:25.283Z
   [junit4]   2> 811490 INFO  (zkConnectionManagerCallback-2769-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811490 INFO  (jetty-launcher-2764-thread-1) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 811495 INFO  (zkConnectionManagerCallback-2773-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811496 INFO  (jetty-launcher-2764-thread-4) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 811514 INFO  (zkConnectionManagerCallback-2775-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811519 INFO  (jetty-launcher-2764-thread-2) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 811540 INFO  (zkConnectionManagerCallback-2771-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811549 INFO  (jetty-launcher-2764-thread-3) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 811732 INFO  (jetty-launcher-2764-thread-1) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44880/solr
   [junit4]   2> 811733 INFO  (jetty-launcher-2764-thread-2) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44880/solr
   [junit4]   2> 811734 INFO  (zkConnectionManagerCallback-2780-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811737 INFO  (zkConnectionManagerCallback-2783-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811737 INFO  (zkConnectionManagerCallback-2785-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811740 INFO  (zkConnectionManagerCallback-2787-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811763 INFO  (jetty-launcher-2764-thread-4) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44880/solr
   [junit4]   2> 811777 INFO  (zkConnectionManagerCallback-2795-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811779 INFO  (jetty-launcher-2764-thread-3) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44880/solr
   [junit4]   2> 811789 INFO  (zkConnectionManagerCallback-2799-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811789 INFO  (zkConnectionManagerCallback-2801-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811791 INFO  (zkConnectionManagerCallback-2803-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811810 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:44893_solr
   [junit4]   2> 811812 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.c.Overseer Overseer (id=72220912799580171-127.0.0.1:44893_solr-n_0000000000) starting
   [junit4]   2> 811812 INFO  (jetty-launcher-2764-thread-3) [n:127.0.0.1:36278_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:36278_solr
   [junit4]   2> 811824 INFO  (zkCallback-2802-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 811824 INFO  (zkCallback-2800-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 811834 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 811834 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 811839 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 811839 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:42205_solr
   [junit4]   2> 811840 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 811840 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:34594_solr
   [junit4]   2> 811857 INFO  (zkConnectionManagerCallback-2812-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811858 INFO  (zkCallback-2800-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 811858 INFO  (zkCallback-2802-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 811858 INFO  (zkConnectionManagerCallback-2817-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811858 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 811859 INFO  (zkCallback-2802-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 811859 INFO  (zkCallback-2800-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (2) -> (3)
   [junit4]   2> 811863 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44880/solr ready
   [junit4]   2> 811863 INFO  (jetty-launcher-2764-thread-3) [n:127.0.0.1:36278_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (3)
   [junit4]   2> 811864 INFO  (jetty-launcher-2764-thread-3) [n:127.0.0.1:36278_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44880/solr ready
   [junit4]   2> 811867 INFO  (zkCallback-2786-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (3)
   [junit4]   2> 811870 INFO  (zkCallback-2784-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (3)
   [junit4]   2> 811877 INFO  (OverseerStateUpdate-72220912799580171-127.0.0.1:44893_solr-n_0000000000) [n:127.0.0.1:44893_solr    ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:44893_solr
   [junit4]   2> 811879 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 811879 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:44893_solr
   [junit4]   2> 811879 INFO  (jetty-launcher-2764-thread-3) [n:127.0.0.1:36278_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 811882 INFO  (zkCallback-2802-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 811885 INFO  (zkConnectionManagerCallback-2824-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811888 INFO  (zkCallback-2800-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 811888 INFO  (zkCallback-2816-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 811889 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
   [junit4]   2> 811889 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44880/solr ready
   [junit4]   2> 811891 INFO  (zkConnectionManagerCallback-2828-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 811891 INFO  (zkCallback-2784-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 811891 INFO  (zkCallback-2811-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 811891 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 811892 INFO  (zkCallback-2786-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (3) -> (4)
   [junit4]   2> 811892 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
   [junit4]   2> 811897 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 811914 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44880/solr ready
   [junit4]   2> 811916 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_44893.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 811918 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 811934 INFO  (jetty-launcher-2764-thread-3) [n:127.0.0.1:36278_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_36278.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 811954 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_44893.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 811955 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_44893.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 811966 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34594.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 812131 INFO  (jetty-launcher-2764-thread-4) [n:127.0.0.1:44893_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node4/.
   [junit4]   2> 811967 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_42205.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 811967 INFO  (jetty-launcher-2764-thread-3) [n:127.0.0.1:36278_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_36278.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 812150 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34594.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 812150 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34594.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 812151 INFO  (jetty-launcher-2764-thread-1) [n:127.0.0.1:34594_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node1/.
   [junit4]   2> 812156 INFO  (jetty-launcher-2764-thread-3) [n:127.0.0.1:36278_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_36278.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 812157 INFO  (jetty-launcher-2764-thread-3) [n:127.0.0.1:36278_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node3/.
   [junit4]   2> 812162 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_42205.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 812163 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_42205.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 812164 INFO  (jetty-launcher-2764-thread-2) [n:127.0.0.1:42205_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node2/.
   [junit4]   2> 812272 INFO  (zkConnectionManagerCallback-2833-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 812280 INFO  (zkConnectionManagerCallback-2838-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 812281 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
   [junit4]   2> 812282 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44880/solr ready
   [junit4]   2> 812286 INFO  (qtp1532940053-10677) [n:127.0.0.1:44893_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :overseerstatus with params action=OVERSEERSTATUS&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 812298 INFO  (qtp1532940053-10677) [n:127.0.0.1:44893_solr    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={action=OVERSEERSTATUS&wt=javabin&version=2} status=0 QTime=11
   [junit4]   1> Formatting using clusterid: testClusterID
   [junit4]   2> 812412 WARN  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.h.m.i.MetricsConfig Cannot locate configuration: tried hadoop-metrics2-namenode.properties,hadoop-metrics2.properties
   [junit4]   2> 812434 WARN  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 812435 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 812456 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/hdfs to ./temp/Jetty_localhost_36102_hdfs____.rabpuu/webapp
   [junit4]   2> 812934 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:36102
   [junit4]   2> 813044 WARN  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 813045 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 813073 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_45939_datanode____.fi4kpn/webapp
   [junit4]   2> 813442 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:45939
   [junit4]   2> 813603 WARN  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.a.h.h.HttpRequestLog Jetty request log can only be enabled using Log4j
   [junit4]   2> 813604 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log jetty-6.1.26
   [junit4]   2> 813619 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log Extract jar:file:/x1/jenkins/.ivy2/cache/org.apache.hadoop/hadoop-hdfs/tests/hadoop-hdfs-2.7.4-tests.jar!/webapps/datanode to ./temp/Jetty_localhost_33846_datanode____vaodvl/webapp
   [junit4]   2> 813808 ERROR (DataNode: [[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-002/hdfsBaseDir/data/data1/, [DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-002/hdfsBaseDir/data/data2/]]  heartbeating to localhost/127.0.0.1:33340) [    ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
   [junit4]   2> 813863 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0x8da8160a26a8c: from storage DS-ea4a4b2b-89a4-4493-a227-3405e1b1f3ce node DatanodeRegistration(127.0.0.1:36983, datanodeUuid=a3338ece-8ba3-4ecf-b028-867219dc2951, infoPort=45282, infoSecurePort=0, ipcPort=41284, storageInfo=lv=-56;cid=testClusterID;nsid=680811697;c=0), blocks: 0, hasStaleStorage: true, processing time: 0 msecs
   [junit4]   2> 813863 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0x8da8160a26a8c: from storage DS-3788ed23-d55b-4c84-8b05-8c9a82281d59 node DatanodeRegistration(127.0.0.1:36983, datanodeUuid=a3338ece-8ba3-4ecf-b028-867219dc2951, infoPort=45282, infoSecurePort=0, ipcPort=41284, storageInfo=lv=-56;cid=testClusterID;nsid=680811697;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 814107 INFO  (SUITE-MoveReplicaHDFSTest-seed#[DDC8408B1482A6D5]-worker) [    ] o.m.log Started HttpServer2$SelectChannelConnectorWithSafeStartup@localhost:33846
   [junit4]   2> 814312 INFO  (OverseerCollectionConfigSetProcessor-72220912799580171-127.0.0.1:44893_solr-n_0000000000) [n:127.0.0.1:44893_solr    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 814346 ERROR (DataNode: [[[DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-002/hdfsBaseDir/data/data3/, [DISK]file:/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-002/hdfsBaseDir/data/data4/]]  heartbeating to localhost/127.0.0.1:33340) [    ] o.a.h.h.s.d.DirectoryScanner dfs.datanode.directoryscan.throttle.limit.ms.per.sec set to value below 1 ms/sec. Assuming default value of 1000
   [junit4]   2> 814367 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0x8da817ec99e2e: from storage DS-cce0230f-ee5b-4faf-b1ed-12578caa2139 node DatanodeRegistration(127.0.0.1:39810, datanodeUuid=4d27a510-24b3-4e98-9182-6bd9d7280764, infoPort=42559, infoSecurePort=0, ipcPort=43163, storageInfo=lv=-56;cid=testClusterID;nsid=680811697;c=0), blocks: 0, hasStaleStorage: true, processing time: 1 msecs
   [junit4]   2> 814367 INFO  (Block report processor) [    ] BlockStateChange BLOCK* processReport 0x8da817ec99e2e: from storage DS-1acb55d3-edcb-4ec7-8f42-e1c88dc5be22 node DatanodeRegistration(127.0.0.1:39810, datanodeUuid=4d27a510-24b3-4e98-9182-6bd9d7280764, infoPort=42559, infoSecurePort=0, ipcPort=43163, storageInfo=lv=-56;cid=testClusterID;nsid=680811697;c=0), blocks: 0, hasStaleStorage: false, processing time: 0 msecs
   [junit4]   2> 814864 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.SolrTestCaseJ4 ###Starting testFailedMove
   [junit4]   2> 814872 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
   [junit4]   2> 814882 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
   [junit4]   2> 814956 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 814956 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 814956 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 814962 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@21976b06{/solr,null,AVAILABLE}
   [junit4]   2> 814962 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.e.j.s.AbstractConnector Started ServerConnector@6f921073{HTTP/1.1,[http/1.1]}{127.0.0.1:36493}
   [junit4]   2> 814962 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.e.j.s.Server Started @815051ms
   [junit4]   2> 814962 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {hostContext=/solr, hostPort=36493}
   [junit4]   2> 814962 ERROR (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 814962 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 814963 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solrâ„¢ version 7.5.0
   [junit4]   2> 814963 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 814963 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 814963 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-08-25T13:21:28.758Z
   [junit4]   2> 814977 INFO  (zkConnectionManagerCallback-2842-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 814978 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.s.SolrDispatchFilter solr.xml found in ZooKeeper. Loading...
   [junit4]   2> 815605 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:44880/solr
   [junit4]   2> 815627 INFO  (zkConnectionManagerCallback-2846-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 815648 INFO  (zkConnectionManagerCallback-2848-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 815654 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (4)
   [junit4]   2> 815657 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 815657 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:36493_solr
   [junit4]   2> 815658 INFO  (zkCallback-2816-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815662 INFO  (zkCallback-2811-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815662 INFO  (zkCallback-2800-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815662 INFO  (zkCallback-2786-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815662 INFO  (zkCallback-2827-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815662 INFO  (zkCallback-2784-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815662 INFO  (zkCallback-2802-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815668 INFO  (zkCallback-2823-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815668 INFO  (zkCallback-2837-thread-2) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815668 INFO  (zkCallback-2837-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815668 INFO  (zkCallback-2847-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (4) -> (5)
   [junit4]   2> 815678 INFO  (zkConnectionManagerCallback-2855-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 815679 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (5)
   [junit4]   2> 815680 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:44880/solr ready
   [junit4]   2> 815680 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 815696 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_36493.solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 815709 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_36493.solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 815709 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_36493.solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 815710 INFO  (TEST-MoveReplicaHDFSTest.testFailedMove-seed#[DDC8408B1482A6D5]) [n:127.0.0.1:36493_solr    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node5/.
   [junit4]   2> 815772 INFO  (zkConnectionManagerCallback-2858-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 815780 INFO  (qtp1582552497-10652) [n:127.0.0.1:34594_solr    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&autoAddReplicas=false&name=MoveReplicaHDFSTest_failed_coll_true&nrtReplicas=2&action=CREATE&numShards=2&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 815792 INFO  (OverseerThreadFactory-4756-thread-2-processing-n:127.0.0.1:44893_solr) [n:127.0.0.1:44893_solr    ] o.a.s.c.a.c.CreateCollectionCmd Create collection MoveReplicaHDFSTest_failed_coll_true
   [junit4]   2> 815902 INFO  (OverseerStateUpdate-72220912799580171-127.0.0.1:44893_solr-n_0000000000) [n:127.0.0.1:44893_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"MoveReplicaHDFSTest_failed_coll_true",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:44893/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 815904 INFO  (OverseerStateUpdate-72220912799580171-127.0.0.1:44893_solr-n_0000000000) [n:127.0.0.1:44893_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"MoveReplicaHDFSTest_failed_coll_true",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "core":"MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:34594/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 815905 INFO  (OverseerStateUpdate-72220912799580171-127.0.0.1:44893_solr-n_0000000000) [n:127.0.0.1:44893_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"MoveReplicaHDFSTest_failed_coll_true",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:36493/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 815907 INFO  (OverseerStateUpdate-72220912799580171-127.0.0.1:44893_solr-n_0000000000) [n:127.0.0.1:44893_solr    ] o.a.s.c.o.SliceMutator createReplica() {
   [junit4]   2>   "operation":"ADDREPLICA",
   [junit4]   2>   "collection":"MoveReplicaHDFSTest_failed_coll_true",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "core":"MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "base_url":"http://127.0.0.1:36278/solr",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "waitForFinalState":"false"} 
   [junit4]   2> 816181 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr    x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node5&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_failed_coll_true&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 816181 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr    x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node3&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_failed_coll_true&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 816186 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr    x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node7&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_failed_coll_true&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 816186 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr    x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&coreNodeName=core_node8&collection.configName=conf1&newCollection=true&name=MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6&action=CREATE&numShards=2&collection=MoveReplicaHDFSTest_failed_coll_true&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 816186 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr    x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 2147483647 transient cores
   [junit4]   2> 817207 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
   [junit4]   2> 817245 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
   [junit4]   2> 817250 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
   [junit4]   2> 817255 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] Schema name=minimal
   [junit4]   2> 817257 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] Schema name=minimal
   [junit4]   2> 817257 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
   [junit4]   2> 817258 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 817258 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2' using configuration from collection MoveReplicaHDFSTest_failed_coll_true, trusted=true
   [junit4]   2> 817258 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 817258 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_34594.solr.core.MoveReplicaHDFSTest_failed_coll_true.shard1.replica_n2' (registry 'solr.core.MoveReplicaHDFSTest_failed_coll_true.shard1.replica_n2') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 817258 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4' using configuration from collection MoveReplicaHDFSTest_failed_coll_true, trusted=true
   [junit4]   2> 817259 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_36493.solr.core.MoveReplicaHDFSTest_failed_coll_true.shard2.replica_n4' (registry 'solr.core.MoveReplicaHDFSTest_failed_coll_true.shard2.replica_n4') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 817268 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] Schema name=minimal
   [junit4]   2> 817268 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.s.IndexSchema [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] Schema name=minimal
   [junit4]   2> 817269 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost:33340/data
   [junit4]   2> 817269 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 817269 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 817269 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node1/MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2], dataDir=[hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node5/data/]
   [junit4]   2> 817270 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 817270 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1' using configuration from collection MoveReplicaHDFSTest_failed_coll_true, trusted=true
   [junit4]   2> 817270 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node5/data/snapshot_metadata
   [junit4]   2> 817271 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_44893.solr.core.MoveReplicaHDFSTest_failed_coll_true.shard1.replica_n1' (registry 'solr.core.MoveReplicaHDFSTest_failed_coll_true.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 817271 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost:33340/data
   [junit4]   2> 817271 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 817271 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 817271 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node4/MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1], dataDir=[hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node3/data/]
   [junit4]   2> 817272 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node3/data/snapshot_metadata
   [junit4]   2> 817273 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.s.IndexSchema Loaded schema minimal/1.1 with uniqueid field id
   [junit4]   2> 817273 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.CoreContainer Creating SolrCore 'MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6' using configuration from collection MoveReplicaHDFSTest_failed_coll_true, trusted=true
   [junit4]   2> 817274 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr_36278.solr.core.MoveReplicaHDFSTest_failed_coll_true.shard2.replica_n6' (registry 'solr.core.MoveReplicaHDFSTest_failed_coll_true.shard2.replica_n6') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@2f7bf3f3
   [junit4]   2> 817274 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost:33340/data
   [junit4]   2> 817274 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 817275 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory solr.hdfs.home=hdfs://localhost:33340/data
   [junit4]   2> 817285 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 817285 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory Solr Kerberos Authentication disabled
   [junit4]   2> 817285 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 817285 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node5/MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4], dataDir=[hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node7/data/]
   [junit4]   2> 817286 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node7/data/snapshot_metadata
   [junit4]   2> 817305 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SolrCore [[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.MoveReplicaHDFSTest_DDC8408B1482A6D5-001/tempDir-001/node3/MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6], dataDir=[hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node8/data/]
   [junit4]   2> 817306 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node8/data/snapshot_metadata
   [junit4]   2> 817352 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node5/data
   [junit4]   2> 817375 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node7/data
   [junit4]   2> 817382 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node3/data
   [junit4]   2> 817382 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node8/data
   [junit4]   2> 817430 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node5/data/index
   [junit4]   2> 817436 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node7/data/index
   [junit4]   2> 817449 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node8/data/index
   [junit4]   2> 817489 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.HdfsDirectoryFactory creating directory factory for path hdfs://localhost:33340/data/MoveReplicaHDFSTest_failed_coll_true/core_node3/data/index
   [junit4]   2> 817810 INFO  (Block report processor) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:39810 is added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-3788ed23-d55b-4c84-8b05-8c9a82281d59:NORMAL:127.0.0.1:36983|RBW], ReplicaUC[[DISK]DS-1acb55d3-edcb-4ec7-8f42-e1c88dc5be22:NORMAL:127.0.0.1:39810|RBW]]} size 0
   [junit4]   2> 817857 INFO  (Block report processor) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:36983 is added to blk_1073741825_1001{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-1acb55d3-edcb-4ec7-8f42-e1c88dc5be22:NORMAL:127.0.0.1:39810|RBW], ReplicaUC[[DISK]DS-ea4a4b2b-89a4-4493-a227-3405e1b1f3ce:NORMAL:127.0.0.1:36983|FINALIZED]]} size 0
   [junit4]   2> 817958 INFO  (Block report processor) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:39810 is added to blk_1073741827_1003{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-3788ed23-d55b-4c84-8b05-8c9a82281d59:NORMAL:127.0.0.1:36983|RBW], ReplicaUC[[DISK]DS-cce0230f-ee5b-4faf-b1ed-12578caa2139:NORMAL:127.0.0.1:39810|RBW]]} size 0
   [junit4]   2> 817962 INFO  (Block report processor) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:36983 is added to blk_1073741827_1003{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-3788ed23-d55b-4c84-8b05-8c9a82281d59:NORMAL:127.0.0.1:36983|RBW], ReplicaUC[[DISK]DS-cce0230f-ee5b-4faf-b1ed-12578caa2139:NORMAL:127.0.0.1:39810|RBW]]} size 0
   [junit4]   2> 818011 INFO  (Block report processor) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:36983 is added to blk_1073741828_1004{UCState=UNDER_CONSTRUCTION, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-3788ed23-d55b-4c84-8b05-8c9a82281d59:NORMAL:127.0.0.1:36983|RBW], ReplicaUC[[DISK]DS-1acb55d3-edcb-4ec7-8f42-e1c88dc5be22:NORMAL:127.0.0.1:39810|RBW]]} size 0
   [junit4]   2> 818016 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 818016 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 818016 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
   [junit4]   2> 818018 INFO  (Block report processor) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:36983 is added to blk_1073741826_1002{UCState=COMMITTED, truncateBlock=null, primaryNodeIndex=-1, replicas=[ReplicaUC[[DISK]DS-ea4a4b2b-89a4-4493-a227-3405e1b1f3ce:NORMAL:127.0.0.1:36983|RBW], ReplicaUC[[DISK]DS-cce0230f-ee5b-4faf-b1ed-12578caa2139:NORMAL:127.0.0.1:39810|RBW]]} size 69
   [junit4]   2> 818019 INFO  (Block report processor) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:39810 is added to blk_1073741828_1004 size 69
   [junit4]   2> 818020 INFO  (Block report processor) [    ] BlockStateChange BLOCK* addStoredBlock: blockMap updated: 127.0.0.1:39810 is added to blk_1073741826_1002 size 69
   [junit4]   2> 818051 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 818051 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 818254 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.s.SolrIndexSearcher Opening [Searcher@1b3bdef0[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] main]
   [junit4]   2> 818255 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 818255 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 818256 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 818256 INFO  (searcherExecutor-4778-thread-1-processing-n:127.0.0.1:34594_solr x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2 c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SolrCore [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] Registered new searcher Searcher@1b3bdef0[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 818256 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1609777327165669376
   [junit4]   2> 818263 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard1 to Terms{values={core_node5=0}, version=0}
   [junit4]   2> 818265 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard1: total=2 found=1 timeoutin=9999ms
   [junit4]   2> 818304 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 818304 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 818304 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
   [junit4]   2> 818326 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 818326 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 818328 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 818328 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 818328 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
   [junit4]   2> 818353 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 818355 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 818398 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@684f4667[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] main]
   [junit4]   2> 818399 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 818400 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 818400 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 818401 INFO  (searcherExecutor-4780-thread-1-processing-n:127.0.0.1:44893_solr x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1 c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.SolrCore [MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] Registered new searcher Searcher@684f4667[MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 818408 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1609777327325052928
   [junit4]   2> 818411 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.s.SolrIndexSearcher Opening [Searcher@73200847[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] main]
   [junit4]   2> 818412 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 818413 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 818413 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 818414 INFO  (searcherExecutor-4781-thread-1-processing-n:127.0.0.1:36278_solr x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6 c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.SolrCore [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] Registered new searcher Searcher@73200847[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 818414 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1609777327331344384
   [junit4]   2> 818420 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard2 to Terms{values={core_node8=0}, version=0}
   [junit4]   2> 818422 INFO  (qtp1532940053-10659) [n:127.0.0.1:44893_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node3 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard1 to Terms{values={core_node3=0, core_node5=0}, version=1}
   [junit4]   2> 818424 INFO  (qtp535344686-10660) [n:127.0.0.1:36278_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node8 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n6] o.a.s.c.ShardLeaderElectionContext Waiting until we see more replicas up for shard shard2: total=2 found=1 timeoutin=9999ms
   [junit4]   2> 818569 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.HdfsUpdateLog
   [junit4]   2> 818569 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir=null defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 818569 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.HdfsUpdateLog Initializing HdfsUpdateLog: tlogDfsReplication=3
   [junit4]   2> 818590 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 818590 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 818639 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.s.SolrIndexSearcher Opening [Searcher@799228b3[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] main]
   [junit4]   2> 818640 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 818640 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 818641 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 818642 INFO  (searcherExecutor-4779-thread-1-processing-n:127.0.0.1:36493_solr x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4 c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.SolrCore [MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] Registered new searcher Searcher@799228b3[MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 818642 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1609777327570419712
   [junit4]   2> 818646 INFO  (qtp1006370055-10952) [n:127.0.0.1:36493_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard2 r:core_node7 x:MoveReplicaHDFSTest_failed_coll_true_shard2_replica_n4] o.a.s.c.ZkShardTerms Successful update of terms at /collections/MoveReplicaHDFSTest_failed_coll_true/terms/shard2 to Terms{values={core_node7=0, core_node8=0}, version=1}
   [junit4]   2> 818766 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 818766 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 818766 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:34594/solr/MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2/
   [junit4]   2> 818767 INFO  (qtp1582552497-10645) [n:127.0.0.1:34594_solr c:MoveReplicaHDFSTest_failed_coll_true s:shard1 r:core_node5 x:MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2] o.a.s.u.PeerSync PeerSync: core=MoveReplicaHDFSTest_failed_coll_true_shard1_replica_n2 url=http://127.0.0.1:34594/

[...truncated too long message...]

eck:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

jar-checksums:
    [mkdir] Created dir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1285808332
     [copy] Copying 239 files to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1285808332
   [delete] Deleting directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1285808332

check-working-copy:
[ivy:cachepath] :: resolving dependencies :: org.eclipse.jgit#org.eclipse.jgit-caller;working
[ivy:cachepath] 	confs: [default]
[ivy:cachepath] 	found org.eclipse.jgit#org.eclipse.jgit;4.6.0.201612231935-r in public
[ivy:cachepath] 	found com.jcraft#jsch;0.1.53 in public
[ivy:cachepath] 	found com.googlecode.javaewah#JavaEWAH;1.1.6 in public
[ivy:cachepath] 	found org.apache.httpcomponents#httpclient;4.3.6 in public
[ivy:cachepath] 	found org.apache.httpcomponents#httpcore;4.3.3 in public
[ivy:cachepath] 	found commons-logging#commons-logging;1.1.3 in public
[ivy:cachepath] 	found commons-codec#commons-codec;1.6 in public
[ivy:cachepath] 	found org.slf4j#slf4j-api;1.7.2 in public
[ivy:cachepath] :: resolution report :: resolve 245ms :: artifacts dl 3ms
	---------------------------------------------------------------------
	|                  |            modules            ||   artifacts   |
	|       conf       | number| search|dwnlded|evicted|| number|dwnlded|
	---------------------------------------------------------------------
	|      default     |   8   |   0   |   0   |   0   ||   8   |   0   |
	---------------------------------------------------------------------
[wc-checker] Initializing working copy...
[wc-checker] SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
[wc-checker] SLF4J: Defaulting to no-operation (NOP) logger implementation
[wc-checker] SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.
[wc-checker] Checking working copy status...

-jenkins-base:

BUILD SUCCESSFUL
Total time: 90 minutes 5 seconds
Archiving artifacts
java.lang.InterruptedException: no matches found within 10000
	at hudson.FilePath$34.hasMatch(FilePath.java:2678)
	at hudson.FilePath$34.invoke(FilePath.java:2557)
	at hudson.FilePath$34.invoke(FilePath.java:2547)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2918)
Also:   hudson.remoting.Channel$CallSiteStackTrace: Remote call to lucene
		at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
		at hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
		at hudson.remoting.Channel.call(Channel.java:955)
		at hudson.FilePath.act(FilePath.java:1036)
		at hudson.FilePath.act(FilePath.java:1025)
		at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
		at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
		at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
		at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
		at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
		at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
		at hudson.model.Build$BuildExecution.post2(Build.java:186)
		at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
		at hudson.model.Run.execute(Run.java:1819)
		at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
		at hudson.model.ResourceController.execute(ResourceController.java:97)
		at hudson.model.Executor.run(Executor.java:429)
Caused: hudson.FilePath$TunneledInterruptedException
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2920)
	at hudson.remoting.UserRequest.perform(UserRequest.java:212)
	at hudson.remoting.UserRequest.perform(UserRequest.java:54)
	at hudson.remoting.Request$2.run(Request.java:369)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:748)
Caused: java.lang.InterruptedException: java.lang.InterruptedException: no matches found within 10000
	at hudson.FilePath.act(FilePath.java:1038)
	at hudson.FilePath.act(FilePath.java:1025)
	at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
	at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
	at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
	at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
	at hudson.model.Build$BuildExecution.post2(Build.java:186)
	at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
	at hudson.model.Run.execute(Run.java:1819)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
No artifacts found that match the file pattern "**/*.events,heapdumps/**,**/hs_err_pid*". Configuration error?
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)

[JENKINS] Lucene-Solr-Tests-7.x - Build # 816 - Still Unstable

Posted by Apache Jenkins Server <je...@builds.apache.org>.
Build: https://builds.apache.org/job/Lucene-Solr-Tests-7.x/816/

1 tests failed.
FAILED:  org.apache.solr.cloud.api.collections.ShardSplitTest.testSplitMixedReplicaTypes

Error Message:
unexpected shard state expected:<inactive> but was:<active>

Stack Trace:
java.lang.AssertionError: unexpected shard state expected:<inactive> but was:<active>
	at __randomizedtesting.SeedInfo.seed([D7FFA526FAA9D050:6F3CF18606720525]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.junit.Assert.failNotEquals(Assert.java:647)
	at org.junit.Assert.assertEquals(Assert.java:128)
	at org.apache.solr.cloud.api.collections.ShardSplitTest.verifyShard(ShardSplitTest.java:372)
	at org.apache.solr.cloud.api.collections.ShardSplitTest.doSplitMixedReplicaTypes(ShardSplitTest.java:364)
	at org.apache.solr.cloud.api.collections.ShardSplitTest.testSplitMixedReplicaTypes(ShardSplitTest.java:331)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:934)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:970)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:984)
	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:1008)
	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:983)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:943)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:829)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:879)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:890)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
	at java.lang.Thread.run(Thread.java:748)




Build Log:
[...truncated 14389 lines...]
   [junit4] Suite: org.apache.solr.cloud.api.collections.ShardSplitTest
   [junit4]   2> Creating dataDir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/init-core-data-001
   [junit4]   2> 1980817 INFO  (SUITE-ShardSplitTest-seed#[D7FFA526FAA9D050]-worker) [    ] o.a.s.BaseDistributedSearchTestCase Setting hostContext system property: /
   [junit4]   2> 1980824 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.ZkTestServer STARTING ZK TEST SERVER
   [junit4]   2> 1980824 INFO  (Thread-4306) [    ] o.a.s.c.ZkTestServer client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 1980824 INFO  (Thread-4306) [    ] o.a.s.c.ZkTestServer Starting server
   [junit4]   2> 1980828 ERROR (Thread-4306) [    ] o.a.z.s.ZooKeeperServer ZKShutdownHandler is not registered, so ZooKeeper server won't take any action on ERROR or SHUTDOWN server state changes
   [junit4]   2> 1980924 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.ZkTestServer start zk server on port:35442
   [junit4]   2> 1980930 INFO  (zkConnectionManagerCallback-6518-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1980934 INFO  (zkConnectionManagerCallback-6520-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1980939 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml
   [junit4]   2> 1980940 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/schema15.xml to /configs/conf1/schema.xml
   [junit4]   2> 1980941 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 1980942 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
   [junit4]   2> 1980943 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
   [junit4]   2> 1980944 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
   [junit4]   2> 1980945 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml
   [junit4]   2> 1980946 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 1980947 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 1980947 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
   [junit4]   2> 1980948 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractZkTestCase put /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
   [junit4]   2> 1980950 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractFullDistribZkTestBase Will use NRT replicas unless explicitly asked otherwise
   [junit4]   2> 1981058 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
   [junit4]   2> 1981058 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1981058 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1981059 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session node0 Scavenging every 600000ms
   [junit4]   2> 1981082 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@5f014035{/,null,AVAILABLE}
   [junit4]   2> 1981083 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.AbstractConnector Started ServerConnector@6c0352db{HTTP/1.1,[http/1.1]}{127.0.0.1:33691}
   [junit4]   2> 1981083 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.Server Started @1981130ms
   [junit4]   2> 1981083 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/tempDir-001/control/data, hostContext=/, hostPort=33691, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/control-001/cores}
   [junit4]   2> 1981083 ERROR (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1981083 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1981083 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 7.5.0
   [junit4]   2> 1981083 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1981083 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1981083 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-08-25T03:11:59.296Z
   [junit4]   2> 1981113 INFO  (zkConnectionManagerCallback-6522-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1981114 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 1981114 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/control-001/solr.xml
   [junit4]   2> 1981142 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1981142 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1981144 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1981264 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:35442/solr
   [junit4]   2> 1981265 INFO  (zkConnectionManagerCallback-6526-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1981268 INFO  (zkConnectionManagerCallback-6528-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1981344 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerElectionContext I am going to be the leader 127.0.0.1:33691_
   [junit4]   2> 1981344 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer Overseer (id=72218516401618948-127.0.0.1:33691_-n_0000000000) starting
   [junit4]   2> 1981351 INFO  (zkConnectionManagerCallback-6535-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1981352 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:35442/solr ready
   [junit4]   2> 1981352 INFO  (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer Starting to work on the main queue : 127.0.0.1:33691_
   [junit4]   2> 1981352 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Process current queue of overseer operations
   [junit4]   2> 1981354 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue. #Running tasks: 0
   [junit4]   2> 1981354 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
   [junit4]   2> 1981354 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
   [junit4]   2> 1981354 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
   [junit4]   2> 1981355 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: []
   [junit4]   2> 1981354 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:33691_
   [junit4]   2> 1981356 INFO  (zkCallback-6534-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1981356 INFO  (zkCallback-6527-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1981373 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1981388 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee
   [junit4]   2> 1981396 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee
   [junit4]   2> 1981397 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee
   [junit4]   2> 1981398 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [n:127.0.0.1:33691_    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/control-001/cores
   [junit4]   2> 1981416 INFO  (zkConnectionManagerCallback-6541-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1981417 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1981417 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:35442/solr ready
   [junit4]   2> 1981418 INFO  (qtp950095423-23501) [n:127.0.0.1:33691_    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:33691_&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1981420 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Got 1 tasks from work-queue : [[org.apache.solr.cloud.OverseerTaskQueue$QueueEvent@2e684ecf]]
   [junit4]   2> 1981420 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Marked task [/overseer/collection-queue-work/qn-0000000000] as running
   [junit4]   2> 1981420 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Overseer Collection Message Handler: Get the message id:/overseer/collection-queue-work/qn-0000000000 message:{
   [junit4]   2>   "name":"control_collection",
   [junit4]   2>   "fromApi":"true",
   [junit4]   2>   "collection.configName":"conf1",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "createNodeSet":"127.0.0.1:33691_",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "stateFormat":"2",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "operation":"create"}
   [junit4]   2> 1981420 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Runner processing /overseer/collection-queue-work/qn-0000000000
   [junit4]   2> 1981420 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue. #Running tasks: 1
   [junit4]   2> 1981420 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.OverseerCollectionMessageHandler OverseerCollectionMessageHandler.processMessage : create , {
   [junit4]   2>   "name":"control_collection",
   [junit4]   2>   "fromApi":"true",
   [junit4]   2>   "collection.configName":"conf1",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "createNodeSet":"127.0.0.1:33691_",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "stateFormat":"2",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "operation":"create"}
   [junit4]   2> 1981420 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningTasks: [/overseer/collection-queue-work/qn-0000000000]
   [junit4]   2> 1981420 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
   [junit4]   2> 1981420 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
   [junit4]   2> 1981420 INFO  (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Create collection control_collection
   [junit4]   2> 1981420 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: [/overseer/collection-queue-work/qn-0000000000]
   [junit4]   2> 1981421 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.OverseerCollectionMessageHandler creating collections conf node /collections/control_collection 
   [junit4]   2> 1981421 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Check for collection zkNode:control_collection
   [junit4]   2> 1981422 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Collection zkNode exists
   [junit4]   2> 1981423 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "name":"control_collection",
   [junit4]   2>   "fromApi":"true",
   [junit4]   2>   "collection.configName":"conf1",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "createNodeSet":"127.0.0.1:33691_",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "stateFormat":"2",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "operation":"create"} current state version: 0
   [junit4]   2> 1981423 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ClusterStateMutator building a new cName: control_collection
   [junit4]   2> 1981424 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ZkStateWriter going to create_collection /collections/control_collection/state.json
   [junit4]   2> 1981523 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.Assign Identify nodes using default
   [junit4]   2> 1981523 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Creating SolrCores for new collection control_collection, shardNames [shard1] , message : {
   [junit4]   2>   "name":"control_collection",
   [junit4]   2>   "fromApi":"true",
   [junit4]   2>   "collection.configName":"conf1",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "createNodeSet":"127.0.0.1:33691_",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "stateFormat":"2",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "operation":"create"}
   [junit4]   2> 1981524 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Creating core control_collection_shard1_replica_n1 as part of shard shard1 of collection control_collection on 127.0.0.1:33691_
   [junit4]   2> 1981525 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_    x:control_collection_shard1_replica_n1] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1981526 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_    x:control_collection_shard1_replica_n1] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1981530 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "core":"control_collection_shard1_replica_n1",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "base_url":"http://127.0.0.1:33691",
   [junit4]   2>   "node_name":"127.0.0.1:33691_",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"state"} current state version: 0
   [junit4]   2> 1981530 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Update state numShards=1 message={
   [junit4]   2>   "core":"control_collection_shard1_replica_n1",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "base_url":"http://127.0.0.1:33691",
   [junit4]   2>   "node_name":"127.0.0.1:33691_",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"state"}
   [junit4]   2> 1981531 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Will update state for replica: core_node2:{"core":"control_collection_shard1_replica_n1","base_url":"http://127.0.0.1:33691","node_name":"127.0.0.1:33691_","state":"down","type":"NRT"}
   [junit4]   2> 1981531 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Collection is now: DocCollection(control_collection//collections/control_collection/state.json/0)={
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "shards":{"shard1":{
   [junit4]   2>       "range":"80000000-7fffffff",
   [junit4]   2>       "state":"active",
   [junit4]   2>       "replicas":{"core_node2":{
   [junit4]   2>           "core":"control_collection_shard1_replica_n1",
   [junit4]   2>           "base_url":"http://127.0.0.1:33691",
   [junit4]   2>           "node_name":"127.0.0.1:33691_",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT"}}}}}
   [junit4]   2> 1981631 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ZkStateWriter going to update_collection /collections/control_collection/state.json version: 0
   [junit4]   2> 1982544 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
   [junit4]   2> 1982555 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema [control_collection_shard1_replica_n1] Schema name=test
   [junit4]   2> 1982642 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 1982656 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.CoreContainer Creating SolrCore 'control_collection_shard1_replica_n1' using configuration from collection control_collection, trusted=true
   [junit4]   2> 1982657 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.control_collection.shard1.replica_n1' (registry 'solr.core.control_collection.shard1.replica_n1') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee
   [junit4]   2> 1982657 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 1982657 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [[control_collection_shard1_replica_n1] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/control-001/cores/control_collection_shard1_replica_n1], dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/control-001/cores/control_collection_shard1_replica_n1/data/]
   [junit4]   2> 1982660 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=25, maxMergedSegmentMB=27.2880859375, floorSegmentMB=1.505859375, forceMergeDeletesPctAllowed=13.558130843588996, segmentsPerTier=14.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.605422457767422, deletesPctAllowed=37.5687316093027
   [junit4]   2> 1982669 WARN  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
   [junit4]   2> 1982710 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 1982710 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 1982711 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 1982711 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 1982712 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=27, maxMergeAtOnceExplicit=27, maxMergedSegmentMB=25.072265625, floorSegmentMB=0.654296875, forceMergeDeletesPctAllowed=1.1520060249821729, segmentsPerTier=22.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, deletesPctAllowed=31.98712972722182
   [junit4]   2> 1982713 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.s.SolrIndexSearcher Opening [Searcher@17ccd48f[control_collection_shard1_replica_n1] main]
   [junit4]   2> 1982714 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 1982714 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 1982715 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 1982716 INFO  (searcherExecutor-9027-thread-1-processing-n:127.0.0.1:33691_ x:control_collection_shard1_replica_n1 c:control_collection s:shard1) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SolrCore [control_collection_shard1_replica_n1] Registered new searcher Searcher@17ccd48f[control_collection_shard1_replica_n1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 1982716 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1609738979564847104
   [junit4]   2> 1982721 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ZkShardTerms Successful update of terms at /collections/control_collection/terms/shard1 to Terms{values={core_node2=0}, version=0}
   [junit4]   2> 1982724 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 1982724 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 1982724 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:33691/control_collection_shard1_replica_n1/
   [junit4]   2> 1982724 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 1982724 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.SyncStrategy http://127.0.0.1:33691/control_collection_shard1_replica_n1/ has no replicas
   [junit4]   2> 1982724 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 1982727 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:33691/control_collection_shard1_replica_n1/ shard1
   [junit4]   2> 1982727 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "operation":"leader",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "base_url":"http://127.0.0.1:33691",
   [junit4]   2>   "core":"control_collection_shard1_replica_n1",
   [junit4]   2>   "state":"active"} current state version: 0
   [junit4]   2> 1982828 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ZkStateWriter going to update_collection /collections/control_collection/state.json version: 1
   [junit4]   2> 1982829 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 1982830 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "core":"control_collection_shard1_replica_n1",
   [junit4]   2>   "core_node_name":"core_node2",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "base_url":"http://127.0.0.1:33691",
   [junit4]   2>   "node_name":"127.0.0.1:33691_",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "state":"active",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"state"} current state version: 0
   [junit4]   2> 1982830 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Update state numShards=1 message={
   [junit4]   2>   "core":"control_collection_shard1_replica_n1",
   [junit4]   2>   "core_node_name":"core_node2",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "base_url":"http://127.0.0.1:33691",
   [junit4]   2>   "node_name":"127.0.0.1:33691_",
   [junit4]   2>   "numShards":"1",
   [junit4]   2>   "state":"active",
   [junit4]   2>   "shard":"shard1",
   [junit4]   2>   "collection":"control_collection",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"state"}
   [junit4]   2> 1982830 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Will update state for replica: core_node2:{"core":"control_collection_shard1_replica_n1","base_url":"http://127.0.0.1:33691","node_name":"127.0.0.1:33691_","state":"active","type":"NRT","leader":"true"}
   [junit4]   2> 1982830 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Collection is now: DocCollection(control_collection//collections/control_collection/state.json/2)={
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "shards":{"shard1":{
   [junit4]   2>       "range":"80000000-7fffffff",
   [junit4]   2>       "state":"active",
   [junit4]   2>       "replicas":{"core_node2":{
   [junit4]   2>           "core":"control_collection_shard1_replica_n1",
   [junit4]   2>           "base_url":"http://127.0.0.1:33691",
   [junit4]   2>           "node_name":"127.0.0.1:33691_",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "leader":"true"}}}}}
   [junit4]   2> 1982831 INFO  (qtp950095423-23496) [n:127.0.0.1:33691_ c:control_collection s:shard1  x:control_collection_shard1_replica_n1] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&newCollection=true&name=control_collection_shard1_replica_n1&action=CREATE&numShards=1&collection=control_collection&shard=shard1&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1305
   [junit4]   2> 1982831 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Finished create command on all shards for collection: control_collection
   [junit4]   2> 1982831 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Completed task:[/overseer/collection-queue-work/qn-0000000000]
   [junit4]   2> 1982832 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Marked task [/overseer/collection-queue-work/qn-0000000000] as completed.
   [junit4]   2> 1982832 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
   [junit4]   2> 1982832 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
   [junit4]   2> 1982832 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor CompletedTasks: [/overseer/collection-queue-work/qn-0000000000]
   [junit4]   2> 1982832 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: [/overseer/collection-queue-work/qn-0000000000]
   [junit4]   2> 1982832 DEBUG (OverseerThreadFactory-9022-thread-1-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Overseer Collection Message Handler: Message id:/overseer/collection-queue-work/qn-0000000000 complete, response:{success={127.0.0.1:33691_={responseHeader={status=0,QTime=1305},core=control_collection_shard1_replica_n1}}}
   [junit4]   2> 1982835 INFO  (qtp950095423-23501) [n:127.0.0.1:33691_    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 1982931 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ZkStateWriter going to update_collection /collections/control_collection/state.json version: 2
   [junit4]   2> 1982931 INFO  (zkCallback-6527-thread-1) [    ] o.a.s.c.c.ZkStateReader A cluster state change: [WatchedEvent state:SyncConnected type:NodeDataChanged path:/collections/control_collection/state.json] for collection [control_collection] has occurred - updating... (live nodes size: [1])
   [junit4]   2> 1983424 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Got 0 tasks from work-queue : [[]]
   [junit4]   2> 1983424 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue. #Running tasks: 0
   [junit4]   2> 1983424 INFO  (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000000 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 1983425 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
   [junit4]   2> 1983425 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
   [junit4]   2> 1983425 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
   [junit4]   2> 1983425 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: []
   [junit4]   2> 1983835 INFO  (qtp950095423-23501) [n:127.0.0.1:33691_    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=control_collection&nrtReplicas=1&action=CREATE&numShards=1&createNodeSet=127.0.0.1:33691_&wt=javabin&version=2} status=0 QTime=2417
   [junit4]   2> 1983840 INFO  (zkConnectionManagerCallback-6546-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1983841 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1983842 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:35442/solr ready
   [junit4]   2> 1983842 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.ChaosMonkey monkey: init - expire sessions:false cause connection loss:false
   [junit4]   2> 1983842 INFO  (qtp950095423-23503) [n:127.0.0.1:33691_    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :create with params collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=1&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1983844 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Got 1 tasks from work-queue : [[org.apache.solr.cloud.OverseerTaskQueue$QueueEvent@2e684ed1]]
   [junit4]   2> 1983844 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Marked task [/overseer/collection-queue-work/qn-0000000002] as running
   [junit4]   2> 1983844 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Overseer Collection Message Handler: Get the message id:/overseer/collection-queue-work/qn-0000000002 message:{
   [junit4]   2>   "name":"collection1",
   [junit4]   2>   "fromApi":"true",
   [junit4]   2>   "collection.configName":"conf1",
   [junit4]   2>   "numShards":"2",
   [junit4]   2>   "createNodeSet":"",
   [junit4]   2>   "stateFormat":"1",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "operation":"create"}
   [junit4]   2> 1983844 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Runner processing /overseer/collection-queue-work/qn-0000000002
   [junit4]   2> 1983844 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.OverseerCollectionMessageHandler OverseerCollectionMessageHandler.processMessage : create , {
   [junit4]   2>   "name":"collection1",
   [junit4]   2>   "fromApi":"true",
   [junit4]   2>   "collection.configName":"conf1",
   [junit4]   2>   "numShards":"2",
   [junit4]   2>   "createNodeSet":"",
   [junit4]   2>   "stateFormat":"1",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "operation":"create"}
   [junit4]   2> 1983844 INFO  (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Create collection collection1
   [junit4]   2> 1983844 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue. #Running tasks: 1
   [junit4]   2> 1983845 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningTasks: [/overseer/collection-queue-work/qn-0000000002]
   [junit4]   2> 1983845 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
   [junit4]   2> 1983845 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
   [junit4]   2> 1983845 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: [/overseer/collection-queue-work/qn-0000000002]
   [junit4]   2> 1983845 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.OverseerCollectionMessageHandler creating collections conf node /collections/collection1 
   [junit4]   2> 1983846 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Check for collection zkNode:collection1
   [junit4]   2> 1983846 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Collection zkNode exists
   [junit4]   2> 1983847 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "name":"collection1",
   [junit4]   2>   "fromApi":"true",
   [junit4]   2>   "collection.configName":"conf1",
   [junit4]   2>   "numShards":"2",
   [junit4]   2>   "createNodeSet":"",
   [junit4]   2>   "stateFormat":"1",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "operation":"create"} current state version: 0
   [junit4]   2> 1983847 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ClusterStateMutator building a new cName: collection1
   [junit4]   2> 1984048 WARN  (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd It is unusual to create a collection (collection1) without cores.
   [junit4]   2> 1984050 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.a.c.CreateCollectionCmd Finished create command for collection: collection1
   [junit4]   2> 1984050 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Completed task:[/overseer/collection-queue-work/qn-0000000002]
   [junit4]   2> 1984051 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Marked task [/overseer/collection-queue-work/qn-0000000002] as completed.
   [junit4]   2> 1984051 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
   [junit4]   2> 1984051 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
   [junit4]   2> 1984051 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor CompletedTasks: [/overseer/collection-queue-work/qn-0000000002]
   [junit4]   2> 1984051 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: [/overseer/collection-queue-work/qn-0000000002]
   [junit4]   2> 1984051 DEBUG (OverseerThreadFactory-9022-thread-2-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Overseer Collection Message Handler: Message id:/overseer/collection-queue-work/qn-0000000002 complete, response:{}
   [junit4]   2> 1984052 INFO  (qtp950095423-23503) [n:127.0.0.1:33691_    ] o.a.s.h.a.CollectionsHandler Wait for new collection to be active for at most 30 seconds. Check all shard replicas
   [junit4]   2> 1984052 INFO  (qtp950095423-23503) [n:127.0.0.1:33691_    ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={collection.configName=conf1&name=collection1&nrtReplicas=1&action=CREATE&numShards=2&createNodeSet=&stateFormat=1&wt=javabin&version=2} status=0 QTime=209
   [junit4]   2> 1984162 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 1 in directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-1-001 of type NRT
   [junit4]   2> 1984163 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
   [junit4]   2> 1984163 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1984163 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1984163 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 1984164 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@6da58d6c{/,null,AVAILABLE}
   [junit4]   2> 1984164 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.AbstractConnector Started ServerConnector@3af57caa{HTTP/1.1,[http/1.1]}{127.0.0.1:35489}
   [junit4]   2> 1984164 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.Server Started @1984212ms
   [junit4]   2> 1984164 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/tempDir-001/jetty1, replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/, hostPort=35489, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-1-001/cores}
   [junit4]   2> 1984164 ERROR (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1984164 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1984165 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 7.5.0
   [junit4]   2> 1984165 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1984165 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1984165 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-08-25T03:12:02.378Z
   [junit4]   2> 1984168 INFO  (zkConnectionManagerCallback-6548-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1984169 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 1984169 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-1-001/solr.xml
   [junit4]   2> 1984172 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1984172 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1984173 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1984273 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:35442/solr
   [junit4]   2> 1984274 INFO  (zkConnectionManagerCallback-6552-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1984280 INFO  (zkConnectionManagerCallback-6554-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1984285 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (1)
   [junit4]   2> 1984289 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1984289 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.ZkController Register node as live in ZooKeeper:/live_nodes/127.0.0.1:35489_
   [junit4]   2> 1984289 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "operation":"downnode",
   [junit4]   2>   "node_name":"127.0.0.1:35489_"} current state version: 1
   [junit4]   2> 1984289 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.NodeMutator DownNode state invoked for node: 127.0.0.1:35489_
   [junit4]   2> 1984294 INFO  (zkCallback-6545-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1984294 INFO  (zkCallback-6534-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1984294 INFO  (zkCallback-6553-thread-1) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1984295 INFO  (zkCallback-6527-thread-2) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (1) -> (2)
   [junit4]   2> 1984309 INFO  (zkConnectionManagerCallback-6561-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1984310 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 1984311 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.s.i.ZkClientClusterStateProvider Cluster at 127.0.0.1:35442/solr ready
   [junit4]   2> 1984311 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.h.a.MetricsHistoryHandler No .system collection, keeping metrics history in memory.
   [junit4]   2> 1984337 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.node' (registry 'solr.node') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee
   [junit4]   2> 1984360 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jvm' (registry 'solr.jvm') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee
   [junit4]   2> 1984360 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.jetty' (registry 'solr.jetty') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee
   [junit4]   2> 1984361 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.CorePropertiesLocator Found 0 core definitions underneath /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-1-001/cores
   [junit4]   2> 1984388 INFO  (qtp950095423-23497) [n:127.0.0.1:33691_    ] o.a.s.h.a.CollectionsHandler Invoked Collection Action :addreplica with params node=127.0.0.1:35489_&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2 and sendToOCPQueue=true
   [junit4]   2> 1984390 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Got 1 tasks from work-queue : [[org.apache.solr.cloud.OverseerTaskQueue$QueueEvent@2e684ed3]]
   [junit4]   2> 1984390 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Marked task [/overseer/collection-queue-work/qn-0000000004] as running
   [junit4]   2> 1984390 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Overseer Collection Message Handler: Get the message id:/overseer/collection-queue-work/qn-0000000004 message:{
   [junit4]   2>   "collection":"collection1",
   [junit4]   2>   "node":"127.0.0.1:35489_",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"addreplica"}
   [junit4]   2> 1984390 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Cleaning up work-queue. #Running tasks: 1
   [junit4]   2> 1984391 INFO  (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskQueue Response ZK path: /overseer/collection-queue-work/qnr-0000000002 doesn't exist.  Requestor may have disconnected from ZooKeeper
   [junit4]   2> 1984391 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor Runner processing /overseer/collection-queue-work/qn-0000000004
   [junit4]   2> 1984391 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningTasks: [/overseer/collection-queue-work/qn-0000000004]
   [junit4]   2> 1984391 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
   [junit4]   2> 1984391 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor CompletedTasks: []
   [junit4]   2> 1984391 DEBUG (OverseerCollectionConfigSetProcessor-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: [/overseer/collection-queue-work/qn-0000000004]
   [junit4]   2> 1984391 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.a.c.OverseerCollectionMessageHandler OverseerCollectionMessageHandler.processMessage : addreplica , {
   [junit4]   2>   "collection":"collection1",
   [junit4]   2>   "node":"127.0.0.1:35489_",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"addreplica"}
   [junit4]   2> 1984391 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.a.c.AddReplicaCmd addReplica() : {
   [junit4]   2>   "collection":"collection1",
   [junit4]   2>   "node":"127.0.0.1:35489_",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"addreplica"}
   [junit4]   2> 1984392 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.a.c.Assign getNodesForNewReplicas() shard: shard2 , replicas : 1 , createNodeSet 127.0.0.1:35489_
   [junit4]   2> 1984393 INFO  (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.a.c.AddReplicaCmd Node Identified 127.0.0.1:35489_ for creating new replica of shard shard2
   [junit4]   2> 1984405 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_    x:collection1_shard2_replica_n41] o.a.s.h.a.CoreAdminOperation core create command qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n41&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT
   [junit4]   2> 1984407 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "core":"collection1_shard2_replica_n41",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "base_url":"http://127.0.0.1:35489",
   [junit4]   2>   "node_name":"127.0.0.1:35489_",
   [junit4]   2>   "numShards":"2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "collection":"collection1",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"state"} current state version: 1
   [junit4]   2> 1984407 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Update state numShards=2 message={
   [junit4]   2>   "core":"collection1_shard2_replica_n41",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "base_url":"http://127.0.0.1:35489",
   [junit4]   2>   "node_name":"127.0.0.1:35489_",
   [junit4]   2>   "numShards":"2",
   [junit4]   2>   "state":"down",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "collection":"collection1",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"state"}
   [junit4]   2> 1984410 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Will update state for replica: core_node42:{"core":"collection1_shard2_replica_n41","base_url":"http://127.0.0.1:35489","node_name":"127.0.0.1:35489_","state":"down","type":"NRT"}
   [junit4]   2> 1984410 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Collection is now: DocCollection(collection1//clusterstate.json/2147483647)={
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "shards":{
   [junit4]   2>     "shard1":{
   [junit4]   2>       "range":"80000000-ffffffff",
   [junit4]   2>       "state":"active",
   [junit4]   2>       "replicas":{}},
   [junit4]   2>     "shard2":{
   [junit4]   2>       "range":"0-7fffffff",
   [junit4]   2>       "state":"active",
   [junit4]   2>       "replicas":{"core_node42":{
   [junit4]   2>           "core":"collection1_shard2_replica_n41",
   [junit4]   2>           "base_url":"http://127.0.0.1:35489",
   [junit4]   2>           "node_name":"127.0.0.1:35489_",
   [junit4]   2>           "state":"down",
   [junit4]   2>           "type":"NRT"}}}}}
   [junit4]   2> 1985420 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SolrConfig Using Lucene MatchVersion: 7.5.0
   [junit4]   2> 1985432 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema [collection1_shard2_replica_n41] Schema name=test
   [junit4]   2> 1985515 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.s.IndexSchema Loaded schema test/1.6 with uniqueid field id
   [junit4]   2> 1985527 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.CoreContainer Creating SolrCore 'collection1_shard2_replica_n41' using configuration from collection collection1, trusted=true
   [junit4]   2> 1985528 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.m.r.SolrJmxReporter JMX monitoring for 'solr.core.collection1.shard2.replica_n41' (registry 'solr.core.collection1.shard2.replica_n41') enabled at server: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee
   [junit4]   2> 1985528 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SolrCore solr.RecoveryStrategy.Builder
   [junit4]   2> 1985528 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SolrCore [[collection1_shard2_replica_n41] ] Opening new SolrCore at [/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-1-001/cores/collection1_shard2_replica_n41], dataDir=[/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/../../../../../../../../../../x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-1-001/cores/collection1_shard2_replica_n41/data/]
   [junit4]   2> 1985531 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=33, maxMergeAtOnceExplicit=25, maxMergedSegmentMB=27.2880859375, floorSegmentMB=1.505859375, forceMergeDeletesPctAllowed=13.558130843588996, segmentsPerTier=14.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.605422457767422, deletesPctAllowed=37.5687316093027
   [junit4]   2> 1985543 WARN  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.RequestHandlers INVALID paramSet a in requestHandler {type = requestHandler,name = /dump,class = DumpRequestHandler,attributes = {initParams=a, name=/dump, class=DumpRequestHandler},args = {defaults={a=A,b=B}}}
   [junit4]   2> 1985583 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.UpdateHandler Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 1985583 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.UpdateLog Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10 numVersionBuckets=65536
   [junit4]   2> 1985584 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.CommitTracker Hard AutoCommit: disabled
   [junit4]   2> 1985584 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.CommitTracker Soft AutoCommit: disabled
   [junit4]   2> 1985585 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.RandomMergePolicy RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=27, maxMergeAtOnceExplicit=27, maxMergedSegmentMB=25.072265625, floorSegmentMB=0.654296875, forceMergeDeletesPctAllowed=1.1520060249821729, segmentsPerTier=22.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0, deletesPctAllowed=31.98712972722182
   [junit4]   2> 1985585 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.s.SolrIndexSearcher Opening [Searcher@1e33948c[collection1_shard2_replica_n41] main]
   [junit4]   2> 1985586 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.r.ManagedResourceStorage Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 1985586 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.r.ManagedResourceStorage Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 1985587 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.h.ReplicationHandler Commits will be reserved for 10000ms.
   [junit4]   2> 1985587 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.u.UpdateLog Could not find max version in index or recent updates, using new clock 1609738982575308800
   [junit4]   2> 1985588 INFO  (searcherExecutor-9041-thread-1-processing-n:127.0.0.1:35489_ x:collection1_shard2_replica_n41 c:collection1 s:shard2) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SolrCore [collection1_shard2_replica_n41] Registered new searcher Searcher@1e33948c[collection1_shard2_replica_n41] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 1985592 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ZkShardTerms Successful update of terms at /collections/collection1/terms/shard2 to Terms{values={core_node42=0}, version=0}
   [junit4]   2> 1985594 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext Enough replicas found to continue.
   [junit4]   2> 1985594 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext I may be the new leader - try and sync
   [junit4]   2> 1985594 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy Sync replicas to http://127.0.0.1:35489/collection1_shard2_replica_n41/
   [junit4]   2> 1985594 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy Sync Success - now sync replicas to me
   [junit4]   2> 1985594 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.SyncStrategy http://127.0.0.1:35489/collection1_shard2_replica_n41/ has no replicas
   [junit4]   2> 1985594 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext Found all replicas participating in election, clear LIR
   [junit4]   2> 1985600 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ShardLeaderElectionContext I am the new leader: http://127.0.0.1:35489/collection1_shard2_replica_n41/ shard2
   [junit4]   2> 1985600 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "operation":"leader",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "collection":"collection1",
   [junit4]   2>   "base_url":"http://127.0.0.1:35489",
   [junit4]   2>   "core":"collection1_shard2_replica_n41",
   [junit4]   2>   "state":"active"} current state version: 2
   [junit4]   2> 1985751 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.c.ZkController I am the leader, no recovery necessary
   [junit4]   2> 1985752 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.Overseer processMessage: queueSize: 1, message = {
   [junit4]   2>   "core":"collection1_shard2_replica_n41",
   [junit4]   2>   "core_node_name":"core_node42",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "base_url":"http://127.0.0.1:35489",
   [junit4]   2>   "node_name":"127.0.0.1:35489_",
   [junit4]   2>   "numShards":"2",
   [junit4]   2>   "state":"active",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "collection":"collection1",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"state"} current state version: 3
   [junit4]   2> 1985752 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Update state numShards=2 message={
   [junit4]   2>   "core":"collection1_shard2_replica_n41",
   [junit4]   2>   "core_node_name":"core_node42",
   [junit4]   2>   "roles":null,
   [junit4]   2>   "base_url":"http://127.0.0.1:35489",
   [junit4]   2>   "node_name":"127.0.0.1:35489_",
   [junit4]   2>   "numShards":"2",
   [junit4]   2>   "state":"active",
   [junit4]   2>   "shard":"shard2",
   [junit4]   2>   "collection":"collection1",
   [junit4]   2>   "type":"NRT",
   [junit4]   2>   "operation":"state"}
   [junit4]   2> 1985752 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Will update state for replica: core_node42:{"core":"collection1_shard2_replica_n41","base_url":"http://127.0.0.1:35489","node_name":"127.0.0.1:35489_","state":"active","type":"NRT","leader":"true"}
   [junit4]   2> 1985752 DEBUG (OverseerStateUpdate-72218516401618948-127.0.0.1:33691_-n_0000000000) [n:127.0.0.1:33691_    ] o.a.s.c.o.ReplicaMutator Collection is now: DocCollection(collection1//clusterstate.json/2147483647)={
   [junit4]   2>   "pullReplicas":"0",
   [junit4]   2>   "replicationFactor":"1",
   [junit4]   2>   "router":{"name":"compositeId"},
   [junit4]   2>   "maxShardsPerNode":"1",
   [junit4]   2>   "autoAddReplicas":"false",
   [junit4]   2>   "nrtReplicas":"1",
   [junit4]   2>   "tlogReplicas":"0",
   [junit4]   2>   "shards":{
   [junit4]   2>     "shard1":{
   [junit4]   2>       "range":"80000000-ffffffff",
   [junit4]   2>       "state":"active",
   [junit4]   2>       "replicas":{}},
   [junit4]   2>     "shard2":{
   [junit4]   2>       "range":"0-7fffffff",
   [junit4]   2>       "state":"active",
   [junit4]   2>       "replicas":{"core_node42":{
   [junit4]   2>           "core":"collection1_shard2_replica_n41",
   [junit4]   2>           "base_url":"http://127.0.0.1:35489",
   [junit4]   2>           "node_name":"127.0.0.1:35489_",
   [junit4]   2>           "state":"active",
   [junit4]   2>           "type":"NRT",
   [junit4]   2>           "leader":"true"}}}}}
   [junit4]   2> 1985753 INFO  (qtp852059541-23557) [n:127.0.0.1:35489_ c:collection1 s:shard2  x:collection1_shard2_replica_n41] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/cores params={qt=/admin/cores&collection.configName=conf1&name=collection1_shard2_replica_n41&action=CREATE&collection=collection1&shard=shard2&wt=javabin&version=2&replicaType=NRT} status=0 QTime=1348
   [junit4]   2> 1985754 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.OverseerTaskProcessor Completed task:[/overseer/collection-queue-work/qn-0000000004]
   [junit4]   2> 1985754 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.OverseerTaskProcessor Marked task [/overseer/collection-queue-work/qn-0000000004] as completed.
   [junit4]   2> 1985754 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.OverseerTaskProcessor RunningTasks: []
   [junit4]   2> 1985754 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.OverseerTaskProcessor BlockedTasks: []
   [junit4]   2> 1985754 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.OverseerTaskProcessor CompletedTasks: [/overseer/collection-queue-work/qn-0000000004]
   [junit4]   2> 1985754 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.OverseerTaskProcessor RunningZKTasks: [/overseer/collection-queue-work/qn-0000000004]
   [junit4]   2> 1985755 DEBUG (OverseerThreadFactory-9022-thread-3-processing-n:127.0.0.1:33691_) [n:127.0.0.1:33691_ c:collection1 s:shard2  ] o.a.s.c.OverseerTaskProcessor Overseer Collection Message Handler: Message id:/overseer/collection-queue-work/qn-0000000004 complete, response:{success={127.0.0.1:35489_={responseHeader={status=0,QTime=1348},core=collection1_shard2_replica_n41}}}
   [junit4]   2> 1985755 INFO  (qtp950095423-23497) [n:127.0.0.1:33691_ c:collection1   ] o.a.s.s.HttpSolrCall [admin] webapp=null path=/admin/collections params={node=127.0.0.1:35489_&action=ADDREPLICA&collection=collection1&shard=shard2&type=NRT&wt=javabin&version=2} status=0 QTime=1366
   [junit4]   2> 1985858 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.AbstractFullDistribZkTestBase create jetty 2 in directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-2-001 of type NRT
   [junit4]   2> 1985859 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.Server jetty-9.4.11.v20180605; built: 2018-06-05T18:24:03.829Z; git: d5fc0523cfa96bfebfbda19606cad384d772f04c; jvm 1.8.0_172-b11
   [junit4]   2> 1985859 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session DefaultSessionIdManager workerName=node0
   [junit4]   2> 1985859 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session No SessionScavenger set, using defaults
   [junit4]   2> 1985859 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.session node0 Scavenging every 660000ms
   [junit4]   2> 1985860 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.h.ContextHandler Started o.e.j.s.ServletContextHandler@77f734e8{/,null,AVAILABLE}
   [junit4]   2> 1985860 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.AbstractConnector Started ServerConnector@6c21256e{HTTP/1.1,[http/1.1]}{127.0.0.1:33313}
   [junit4]   2> 1985860 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.e.j.s.Server Started @1985908ms
   [junit4]   2> 1985860 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.s.e.JettySolrRunner Jetty properties: {solr.data.dir=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/tempDir-001/jetty2, replicaType=NRT, solrconfig=solrconfig.xml, hostContext=/, hostPort=33313, coreRootDirectory=/x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-2-001/cores}
   [junit4]   2> 1985860 ERROR (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.u.StartupLoggingUtils Missing Java Option solr.log.dir. Logging may be missing or incomplete.
   [junit4]   2> 1985863 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter Using logger factory org.apache.logging.slf4j.Log4jLoggerFactory
   [junit4]   2> 1985864 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter  ___      _       Welcome to Apache Solr? version 7.5.0
   [junit4]   2> 1985864 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter / __| ___| |_ _   Starting in cloud mode on port null
   [junit4]   2> 1985864 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter \__ \/ _ \ | '_|  Install dir: null
   [junit4]   2> 1985864 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter |___/\___/_|_|    Start time: 2018-08-25T03:12:04.077Z
   [junit4]   2> 1985865 INFO  (zkConnectionManagerCallback-6564-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1985866 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.s.SolrDispatchFilter Loading solr.xml from SolrHome (not found in ZooKeeper)
   [junit4]   2> 1985866 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Loading container configuration from /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/build/solr-core/test/J2/temp/solr.cloud.api.collections.ShardSplitTest_D7FFA526FAA9D050-001/shard-2-001/solr.xml
   [junit4]   2> 1985869 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverWorkLoopDelay is ignored
   [junit4]   2> 1985869 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig Configuration parameter autoReplicaFailoverBadNodeExpiration is ignored
   [junit4]   2> 1985870 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.SolrXmlConfig MBean server found: com.sun.jmx.mbeanserver.JmxMBeanServer@3928eaee, but no JMX reporters were configured - adding default JMX reporter.
   [junit4]   2> 1986022 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.ZkContainer Zookeeper client=127.0.0.1:35442/solr
   [junit4]   2> 1986024 INFO  (zkConnectionManagerCallback-6568-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1986025 INFO  (zkConnectionManagerCallback-6570-thread-1) [    ] o.a.s.c.c.ConnectionManager zkClient has connected
   [junit4]   2> 1986030 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.c.ZkStateReader Updated live nodes from ZooKeeper... (0) -> (2)
   [junit4]   2> 1986033 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTypesLink-seed#[D7FFA526FAA9D050]) [    ] o.a.s.c.TransientSolrCoreCacheDefault Allocating transient cache for 4 transient cores
   [junit4]   2> 1986033 INFO  (TEST-ShardSplitTest.testSplitMixedReplicaTy

[...truncated too long message...]

eck:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

ivy-availability-check:
[loadresource] Do not set property disallowed.ivy.jars.list as its length is 0.

-ivy-fail-disallowed-ivy-version:

ivy-fail:

ivy-fail:

ivy-configure:
[ivy:configure] :: loading settings :: file = /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/lucene/top-level-ivy-settings.xml

resolve:

jar-checksums:
    [mkdir] Created dir: /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1686151715
     [copy] Copying 239 files to /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1686151715
   [delete] Deleting directory /x1/jenkins/jenkins-slave/workspace/Lucene-Solr-Tests-7.x/solr/null1686151715

check-working-copy:
[ivy:cachepath] :: resolving dependencies :: org.eclipse.jgit#org.eclipse.jgit-caller;working
[ivy:cachepath] 	confs: [default]
[ivy:cachepath] 	found org.eclipse.jgit#org.eclipse.jgit;4.6.0.201612231935-r in public
[ivy:cachepath] 	found com.jcraft#jsch;0.1.53 in public
[ivy:cachepath] 	found com.googlecode.javaewah#JavaEWAH;1.1.6 in public
[ivy:cachepath] 	found org.apache.httpcomponents#httpclient;4.3.6 in public
[ivy:cachepath] 	found org.apache.httpcomponents#httpcore;4.3.3 in public
[ivy:cachepath] 	found commons-logging#commons-logging;1.1.3 in public
[ivy:cachepath] 	found commons-codec#commons-codec;1.6 in public
[ivy:cachepath] 	found org.slf4j#slf4j-api;1.7.2 in public
[ivy:cachepath] :: resolution report :: resolve 19ms :: artifacts dl 1ms
	---------------------------------------------------------------------
	|                  |            modules            ||   artifacts   |
	|       conf       | number| search|dwnlded|evicted|| number|dwnlded|
	---------------------------------------------------------------------
	|      default     |   8   |   0   |   0   |   0   ||   8   |   0   |
	---------------------------------------------------------------------
[wc-checker] Initializing working copy...
[wc-checker] SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
[wc-checker] SLF4J: Defaulting to no-operation (NOP) logger implementation
[wc-checker] SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details.
[wc-checker] Checking working copy status...

-jenkins-base:

BUILD SUCCESSFUL
Total time: 93 minutes 12 seconds
Archiving artifacts
java.lang.InterruptedException: no matches found within 10000
	at hudson.FilePath$34.hasMatch(FilePath.java:2678)
	at hudson.FilePath$34.invoke(FilePath.java:2557)
	at hudson.FilePath$34.invoke(FilePath.java:2547)
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2918)
Also:   hudson.remoting.Channel$CallSiteStackTrace: Remote call to lucene
		at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1741)
		at hudson.remoting.UserRequest$ExceptionResponse.retrieve(UserRequest.java:357)
		at hudson.remoting.Channel.call(Channel.java:955)
		at hudson.FilePath.act(FilePath.java:1036)
		at hudson.FilePath.act(FilePath.java:1025)
		at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
		at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
		at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
		at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
		at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
		at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
		at hudson.model.Build$BuildExecution.post2(Build.java:186)
		at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
		at hudson.model.Run.execute(Run.java:1819)
		at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
		at hudson.model.ResourceController.execute(ResourceController.java:97)
		at hudson.model.Executor.run(Executor.java:429)
Caused: hudson.FilePath$TunneledInterruptedException
	at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2920)
	at hudson.remoting.UserRequest.perform(UserRequest.java:212)
	at hudson.remoting.UserRequest.perform(UserRequest.java:54)
	at hudson.remoting.Request$2.run(Request.java:369)
	at hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:72)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
	at java.lang.Thread.run(Thread.java:748)
Caused: java.lang.InterruptedException: java.lang.InterruptedException: no matches found within 10000
	at hudson.FilePath.act(FilePath.java:1038)
	at hudson.FilePath.act(FilePath.java:1025)
	at hudson.FilePath.validateAntFileMask(FilePath.java:2547)
	at hudson.tasks.ArtifactArchiver.perform(ArtifactArchiver.java:243)
	at hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
	at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
	at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:744)
	at hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:690)
	at hudson.model.Build$BuildExecution.post2(Build.java:186)
	at hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:635)
	at hudson.model.Run.execute(Run.java:1819)
	at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43)
	at hudson.model.ResourceController.execute(ResourceController.java:97)
	at hudson.model.Executor.run(Executor.java:429)
No artifacts found that match the file pattern "**/*.events,heapdumps/**,**/hs_err_pid*". Configuration error?
Recording test results
Build step 'Publish JUnit test result report' changed build result to UNSTABLE
Email was triggered for: Unstable (Test Failures)
Sending email for trigger: Unstable (Test Failures)