You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2013/07/14 13:40:06 UTC

[JENKINS] Lucene-Solr-trunk-MacOSX (64bit/jdk1.7.0) - Build # 647 - Still Failing!

Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/647/
Java: 64bit/jdk1.7.0 -XX:+UseCompressedOops -XX:+UseG1GC

1 tests failed.
REGRESSION:  org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testDistribSearch

Error Message:
collection already exists: awholynewcollection_0

Stack Trace:
org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: collection already exists: awholynewcollection_0
	at __randomizedtesting.SeedInfo.seed([DFF894EA9CA91D54:5E1E1AF2EBF67D68]:0)
	at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:424)
	at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)
	at org.apache.solr.client.solrj.impl.LBHttpSolrServer.request(LBHttpSolrServer.java:264)
	at org.apache.solr.client.solrj.impl.CloudSolrServer.request(CloudSolrServer.java:318)
	at org.apache.solr.cloud.AbstractFullDistribZkTestBase.createCollection(AbstractFullDistribZkTestBase.java:1522)
	at org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCollectionsAPI(CollectionsAPIDistributedZkTest.java:438)
	at org.apache.solr.cloud.CollectionsAPIDistributedZkTest.doTest(CollectionsAPIDistributedZkTest.java:146)
	at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:835)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
	at org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
	at java.lang.Thread.run(Thread.java:724)




Build Log:
[...truncated 9662 lines...]
   [junit4] Suite: org.apache.solr.cloud.CollectionsAPIDistributedZkTest
   [junit4]   2> 2507432 T4734 oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system property: /i/v
   [junit4]   2> 2507454 T4734 oas.SolrTestCaseJ4.setUp ###Starting testDistribSearch
   [junit4]   2> Creating dataDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./solrtest-CollectionsAPIDistributedZkTest-1373800440023
   [junit4]   2> 2507457 T4734 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   2> 2507459 T4735 oasc.ZkTestServer$ZKServerMain.runFromConfig Starting server
   [junit4]   2> 2507560 T4734 oasc.ZkTestServer.run start zk server on port:50796
   [junit4]   2> 2507566 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2507578 T4741 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@6cb62e4 name:ZooKeeperConnection Watcher:127.0.0.1:50796 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2507579 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2507581 T4734 oascc.SolrZkClient.makePath makePath: /solr
   [junit4]   2> 2507606 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2507620 T4743 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@45927ba6 name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2507621 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2507622 T4734 oascc.SolrZkClient.makePath makePath: /collections/collection1
   [junit4]   2> 2507636 T4734 oascc.SolrZkClient.makePath makePath: /collections/collection1/shards
   [junit4]   2> 2507652 T4734 oascc.SolrZkClient.makePath makePath: /collections/control_collection
   [junit4]   2> 2507661 T4734 oascc.SolrZkClient.makePath makePath: /collections/control_collection/shards
   [junit4]   2> 2507672 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml
   [junit4]   2> 2507675 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/solrconfig.xml
   [junit4]   2> 2507706 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/schema.xml to /configs/conf1/schema.xml
   [junit4]   2> 2507709 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/schema.xml
   [junit4]   2> 2507723 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
   [junit4]   2> 2507725 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/stopwords.txt
   [junit4]   2> 2507736 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
   [junit4]   2> 2507738 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/protwords.txt
   [junit4]   2> 2507749 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
   [junit4]   2> 2507752 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/currency.xml
   [junit4]   2> 2507763 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 2507765 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/open-exchange-rates.json
   [junit4]   2> 2507773 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 2507775 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 2507785 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
   [junit4]   2> 2507788 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/old_synonyms.txt
   [junit4]   2> 2507804 T4734 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
   [junit4]   2> 2507806 T4734 oascc.SolrZkClient.makePath makePath: /configs/conf1/synonyms.txt
   [junit4]   2> 2508620 T4734 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 2508629 T4734 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50799
   [junit4]   2> 2508640 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 2508641 T4734 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 2508641 T4734 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389
   [junit4]   2> 2508642 T4734 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/'
   [junit4]   2> 2508710 T4734 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/solr.xml
   [junit4]   2> 2508844 T4734 oasc.ConfigSolrXml.<init> Config-defined core root directory: 
   [junit4]   2> 2508846 T4734 oasc.CoreContainer.<init> New CoreContainer 777758190
   [junit4]   2> 2508852 T4734 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/]
   [junit4]   2> 2508859 T4734 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 2508860 T4734 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 2508862 T4734 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 2508863 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 2508864 T4734 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 2508865 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 2508867 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 2508868 T4734 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 2508869 T4734 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 2508870 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 2508899 T4734 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 2508901 T4734 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50796/solr
   [junit4]   2> 2508901 T4734 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 2508903 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2508908 T4754 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@5f54d0d1 name:ZooKeeperConnection Watcher:127.0.0.1:50796 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2508909 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2508915 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 2509033 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2509039 T4756 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@2fdacecd name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2509040 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2509044 T4734 oascc.SolrZkClient.makePath makePath: /overseer/queue
   [junit4]   2> 2509055 T4734 oascc.SolrZkClient.makePath makePath: /overseer/collection-queue-work
   [junit4]   2> 2509063 T4734 oascc.SolrZkClient.makePath makePath: /live_nodes
   [junit4]   2> 2509070 T4734 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50799_i%2Fv
   [junit4]   2> 2509073 T4734 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50799_i%2Fv
   [junit4]   2> 2509083 T4734 oascc.SolrZkClient.makePath makePath: /overseer_elect/election
   [junit4]   2> 2509094 T4734 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 2509101 T4734 oasc.Overseer.start Overseer (id=90033385638330371-127.0.0.1:50799_i%2Fv-n_0000000000) starting
   [junit4]   2> 2509109 T4734 oascc.SolrZkClient.makePath makePath: /overseer/queue-work
   [junit4]   2> 2509128 T4758 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 2509130 T4734 oascc.SolrZkClient.makePath makePath: /clusterstate.json
   [junit4]   2> 2509137 T4734 oascc.SolrZkClient.makePath makePath: /aliases.json
   [junit4]   2> 2509142 T4734 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 2509150 T4757 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 2509157 T4734 oasc.CorePropertiesLocator.discover Looking for core definitions underneath /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389
   [junit4]   2> 2509366 T4734 oasc.CorePropertiesLocator.discoverUnder Found core conf in /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/conf/
   [junit4]   2> 2509378 T4734 oasc.CorePropertiesLocator.discover Found 1 core definitions
   [junit4]   2> 2509390 T4734 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 2509391 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 2509392 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2509418 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:
   [junit4]   2> 2509422 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2509428 T4761 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@3e01177d name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2509428 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2509432 T4734 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 2509440 T4734 oasc.ChaosMonkey.monkeyLog monkey: init - expire sessions:true cause connection loss:true
   [junit4]   2> 2510026 T4734 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 2510034 T4734 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50803
   [junit4]   2> 2510040 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 2510042 T4734 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 2510043 T4734 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008
   [junit4]   2> 2510045 T4734 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/'
   [junit4]   2> 2510127 T4734 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/solr.xml
   [junit4]   2> 2510280 T4734 oasc.ConfigSolrXml.<init> Config-defined core root directory: 
   [junit4]   2> 2510281 T4734 oasc.CoreContainer.<init> New CoreContainer 2042028234
   [junit4]   2> 2510282 T4734 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/]
   [junit4]   2> 2510285 T4734 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 2510286 T4734 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 2510287 T4734 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 2510287 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 2510288 T4734 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 2510289 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 2510289 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 2510290 T4734 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 2510290 T4734 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 2510291 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 2510312 T4734 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 2510313 T4734 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50796/solr
   [junit4]   2> 2510314 T4734 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 2510316 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2510320 T4772 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@576a13a3 name:ZooKeeperConnection Watcher:127.0.0.1:50796 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2510321 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2510324 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 2510343 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2510353 T4774 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@672c7abd name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2510359 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2510370 T4734 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 2511383 T4734 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50803_i%2Fv
   [junit4]   2> 2511387 T4734 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50803_i%2Fv
   [junit4]   2> 2511396 T4756 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 2511397 T4774 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 2511398 T4761 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 2511417 T4734 oasc.CorePropertiesLocator.discover Looking for core definitions underneath /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008
   [junit4]   2> 2511591 T4734 oasc.CorePropertiesLocator.discoverUnder Found core conf in /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/conf/
   [junit4]   2> 2511598 T4734 oasc.CorePropertiesLocator.discover Found 1 core definitions
   [junit4]   2> 2511600 T4734 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 2511601 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 2511601 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2512291 T4734 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 2512319 T4734 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50806
   [junit4]   2> 2512320 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 2512321 T4734 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 2512321 T4734 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194
   [junit4]   2> 2512322 T4734 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/'
   [junit4]   2> 2512393 T4734 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/solr.xml
   [junit4]   2> 2512553 T4734 oasc.ConfigSolrXml.<init> Config-defined core root directory: 
   [junit4]   2> 2512554 T4734 oasc.CoreContainer.<init> New CoreContainer 1444740563
   [junit4]   2> 2512555 T4734 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/]
   [junit4]   2> 2512558 T4734 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 2512559 T4734 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 2512561 T4734 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 2512562 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 2512562 T4734 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 2512563 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 2512564 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 2512569 T4734 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 2512570 T4734 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 2512571 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 2512595 T4734 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 2512597 T4734 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50796/solr
   [junit4]   2> 2512598 T4734 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 2512603 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2512672 T4786 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@4c033bf5 name:ZooKeeperConnection Watcher:127.0.0.1:50796 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2512673 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2512683 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 2512707 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2512711 T4788 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@36b9b998 name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2512712 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2512728 T4734 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 2513737 T4734 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50806_i%2Fv
   [junit4]   2> 2513740 T4734 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50806_i%2Fv
   [junit4]   2> 2513749 T4788 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 2513750 T4761 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 2513750 T4756 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 2513750 T4774 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 2513765 T4734 oasc.CorePropertiesLocator.discover Looking for core definitions underneath /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194
   [junit4]   2> 2513946 T4734 oasc.CorePropertiesLocator.discoverUnder Found core conf in /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/conf/
   [junit4]   2> 2513964 T4734 oasc.CorePropertiesLocator.discover Found 1 core definitions
   [junit4]   2> 2513966 T4734 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 2513967 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 2513971 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2514569 T4734 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 2514577 T4734 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50809
   [junit4]   2> 2514586 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 2514591 T4734 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 2514592 T4734 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558
   [junit4]   2> 2514593 T4734 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/'
   [junit4]   2> 2514638 T4734 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/solr.xml
   [junit4]   2> 2514804 T4734 oasc.ConfigSolrXml.<init> Config-defined core root directory: 
   [junit4]   2> 2514805 T4734 oasc.CoreContainer.<init> New CoreContainer 850965603
   [junit4]   2> 2514806 T4734 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/]
   [junit4]   2> 2514809 T4734 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 2514811 T4734 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 2514811 T4734 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 2514813 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 2514814 T4734 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 2514815 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 2514816 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 2514817 T4734 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 2514819 T4734 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 2514820 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 2514855 T4734 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 2514856 T4734 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50796/solr
   [junit4]   2> 2514857 T4734 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 2514861 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2514866 T4800 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@50dbbbc5 name:ZooKeeperConnection Watcher:127.0.0.1:50796 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2514867 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2514874 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 2514892 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2514899 T4802 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@70e4d902 name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2514899 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2514908 T4734 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 2515917 T4734 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50809_i%2Fv
   [junit4]   2> 2515921 T4734 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50809_i%2Fv
   [junit4]   2> 2515935 T4756 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2515936 T4761 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2515937 T4802 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2515949 T4788 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2515952 T4774 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2515959 T4734 oasc.CorePropertiesLocator.discover Looking for core definitions underneath /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558
   [junit4]   2> 2516147 T4734 oasc.CorePropertiesLocator.discoverUnder Found core conf in /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/conf/
   [junit4]   2> 2516154 T4734 oasc.CorePropertiesLocator.discover Found 1 core definitions
   [junit4]   2> 2516156 T4734 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 2516156 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 2516157 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2516751 T4734 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 2516762 T4734 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50812
   [junit4]   2> 2516764 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 2516765 T4734 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 2516765 T4734 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749
   [junit4]   2> 2516766 T4734 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/'
   [junit4]   2> 2516827 T4734 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/solr.xml
   [junit4]   2> 2516970 T4734 oasc.ConfigSolrXml.<init> Config-defined core root directory: 
   [junit4]   2> 2516971 T4734 oasc.CoreContainer.<init> New CoreContainer 1714793150
   [junit4]   2> 2516972 T4734 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/]
   [junit4]   2> 2516975 T4734 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 2516976 T4734 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 2516977 T4734 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 2516978 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 2516978 T4734 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 2516979 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 2516980 T4734 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 2516980 T4734 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 2516981 T4734 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 2516981 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 2517003 T4734 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 2517005 T4734 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50796/solr
   [junit4]   2> 2517006 T4734 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 2517008 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2517013 T4814 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@4d424ddc name:ZooKeeperConnection Watcher:127.0.0.1:50796 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2517014 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2517017 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 2517045 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2517051 T4816 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@2821dfee name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2517051 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2517060 T4734 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 2518070 T4734 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50812_i%2Fv
   [junit4]   2> 2518074 T4734 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50812_i%2Fv
   [junit4]   2> 2518086 T4802 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 2518087 T4761 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 2518087 T4816 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 2518088 T4788 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 2518099 T4734 oasc.CorePropertiesLocator.discover Looking for core definitions underneath /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749
   [junit4]   2> 2518126 T4774 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 2518127 T4756 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 2518268 T4734 oasc.CorePropertiesLocator.discoverUnder Found core conf in /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/conf/
   [junit4]   2> 2518277 T4734 oasc.CorePropertiesLocator.discover Found 1 core definitions
   [junit4]   2> 2518278 T4734 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 2518278 T4734 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 2518279 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2518305 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2518335 T4765 oasha.CollectionsHandler.handleCreateAction Creating Collection : numShards=2&name=nodes_used_collection&replicationFactor=2&action=CREATE&wt=javabin&version=2
   [junit4]   2> 2518341 T4756 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work state: SyncConnected type NodeChildrenChanged
   [junit4]   2> 2518345 T4758 oasc.OverseerCollectionProcessor.run Overseer Collection Processor: Get the message id:/overseer/collection-queue-work/qn-0000000000 message:{
   [junit4]   2> 	  "operation":"createcollection",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "maxShardsPerNode":null,
   [junit4]   2> 	  "createNodeSet":null,
   [junit4]   2> 	  "name":"nodes_used_collection",
   [junit4]   2> 	  "replicationFactor":"2"}
   [junit4]   2> 2518345 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard nodes_used_collection_shard1_replica1 as part of slice shard1 of collection nodes_used_collection on 127.0.0.1:50806_i%2Fv
   [junit4]   2> 2518346 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard nodes_used_collection_shard1_replica2 as part of slice shard1 of collection nodes_used_collection on 127.0.0.1:50812_i%2Fv
   [junit4]   2> 2518347 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard nodes_used_collection_shard2_replica1 as part of slice shard2 of collection nodes_used_collection on 127.0.0.1:50803_i%2Fv
   [junit4]   2> 2518348 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard nodes_used_collection_shard2_replica2 as part of slice shard2 of collection nodes_used_collection on 127.0.0.1:50809_i%2Fv
   [junit4]   2> 2518367 T4766 oasc.ZkController.publish publishing core=nodes_used_collection_shard2_replica1 state=down
   [junit4]   2> 2518375 T4782 oasc.ZkController.publish publishing core=nodes_used_collection_shard1_replica1 state=down
   [junit4]   2> 2518377 T4782 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2518380 T4811 oasc.ZkController.publish publishing core=nodes_used_collection_shard1_replica2 state=down
   [junit4]   2> 2518383 T4811 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2518385 T4796 oasc.ZkController.publish publishing core=nodes_used_collection_shard2_replica2 state=down
   [junit4]   2> 2518386 T4766 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2518389 T4796 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2519700 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2519702 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for nodes_used_collection, skipping publish state
   [junit4]   2> 2519703 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2519704 T4757 oasc.Overseer$ClusterStateUpdater.createCollection Create collection nodes_used_collection with numShards 2
   [junit4]   2> 2519715 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for nodes_used_collection, skipping publish state
   [junit4]   2> 2519715 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard1_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2519731 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for nodes_used_collection, skipping publish state
   [junit4]   2> 2519732 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard1_replica2",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2519744 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for nodes_used_collection, skipping publish state
   [junit4]   2> 2519745 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica2",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2519760 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2519764 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2519765 T4761 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2519765 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2519766 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2519766 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2520380 T4782 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for nodes_used_collection_shard1_replica1
   [junit4]   2> 2520382 T4782 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/nodes_used_collection_shard1_replica1/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/nodes_used_collection_shard1_replica1/core.properties (No such file or directory)
   [junit4]   2> 2520383 T4782 oasc.CoreContainer.create Creating SolrCore 'nodes_used_collection_shard1_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/nodes_used_collection_shard1_replica1
   [junit4]   2> 2520383 T4782 oasc.ZkController.createCollectionZkNode Check for collection zkNode:nodes_used_collection
   [junit4]   2> 2520385 T4782 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:nodes_used_collection
   [junit4]   2> 2520386 T4782 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2520387 T4811 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for nodes_used_collection_shard1_replica2
   [junit4]   2> 2520388 T4811 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/nodes_used_collection_shard1_replica2/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/nodes_used_collection_shard1_replica2/core.properties (No such file or directory)
   [junit4]   2> 2520389 T4811 oasc.CoreContainer.create Creating SolrCore 'nodes_used_collection_shard1_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/nodes_used_collection_shard1_replica2
   [junit4]   2> 2520389 T4811 oasc.ZkController.createCollectionZkNode Check for collection zkNode:nodes_used_collection
   [junit4]   2> 2520390 T4766 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for nodes_used_collection_shard2_replica1
   [junit4]   2> 2520392 T4766 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/nodes_used_collection_shard2_replica1/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/nodes_used_collection_shard2_replica1/core.properties (No such file or directory)
   [junit4]   2> 2520393 T4766 oasc.CoreContainer.create Creating SolrCore 'nodes_used_collection_shard2_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/nodes_used_collection_shard2_replica1
   [junit4]   2> 2520393 T4766 oasc.ZkController.createCollectionZkNode Check for collection zkNode:nodes_used_collection
   [junit4]   2> 2520391 T4796 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for nodes_used_collection_shard2_replica2
   [junit4]   2> 2520401 T4796 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/nodes_used_collection_shard2_replica2/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/nodes_used_collection_shard2_replica2/core.properties (No such file or directory)
   [junit4]   2> 2520402 T4796 oasc.CoreContainer.create Creating SolrCore 'nodes_used_collection_shard2_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/nodes_used_collection_shard2_replica2
   [junit4]   2> 2520403 T4796 oasc.ZkController.createCollectionZkNode Check for collection zkNode:nodes_used_collection
   [junit4]   2> 2520404 T4811 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:nodes_used_collection
   [junit4]   2> 2520405 T4811 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2520405 T4766 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:nodes_used_collection
   [junit4]   2> 2520406 T4766 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2520405 T4782 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2520408 T4782 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection
   [junit4]   2> 2520407 T4796 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:nodes_used_collection
   [junit4]   2> 2520410 T4796 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2520411 T4766 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2520411 T4766 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection
   [junit4]   2> 2520414 T4811 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2520414 T4811 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection
   [junit4]   2> 2520420 T4796 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2520422 T4796 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection
   [junit4]   2> 2520436 T4766 oasc.ZkController.readConfigName Load collection config from:/collections/nodes_used_collection
   [junit4]   2> 2520437 T4811 oasc.ZkController.readConfigName Load collection config from:/collections/nodes_used_collection
   [junit4]   2> 2520439 T4782 oasc.ZkController.readConfigName Load collection config from:/collections/nodes_used_collection
   [junit4]   2> 2520446 T4796 oasc.ZkController.readConfigName Load collection config from:/collections/nodes_used_collection
   [junit4]   2> 2520446 T4766 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/nodes_used_collection_shard2_replica1/'
   [junit4]   2> 2520447 T4811 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/nodes_used_collection_shard1_replica2/'
   [junit4]   2> 2520452 T4782 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/nodes_used_collection_shard1_replica1/'
   [junit4]   2> 2520455 T4796 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/nodes_used_collection_shard2_replica2/'
   [junit4]   2> 2520670 T4782 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2520728 T4766 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2520753 T4811 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2520769 T4796 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2520973 T4782 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2520995 T4782 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2521013 T4782 oass.IndexSchema.readSchema [nodes_used_collection_shard1_replica1] Schema name=test
   [junit4]   2> 2521031 T4811 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2521060 T4811 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2521075 T4811 oass.IndexSchema.readSchema [nodes_used_collection_shard1_replica2] Schema name=test
   [junit4]   2> 2521090 T4766 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2521095 T4796 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2521103 T4766 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2521118 T4796 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2521139 T4766 oass.IndexSchema.readSchema [nodes_used_collection_shard2_replica1] Schema name=test
   [junit4]   2> 2521162 T4796 oass.IndexSchema.readSchema [nodes_used_collection_shard2_replica2] Schema name=test
   [junit4]   2> 2523937 T4811 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2523970 T4811 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2523993 T4811 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2524003 T4782 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2524033 T4782 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2524041 T4766 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2524043 T4782 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2524062 T4796 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2524066 T4766 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2524082 T4766 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2524088 T4796 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2524094 T4796 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2524121 T4811 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2524144 T4811 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2524158 T4811 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2524163 T4782 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2524168 T4811 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2524169 T4811 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2524170 T4811 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2524180 T4782 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2524183 T4811 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2524184 T4811 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2524185 T4811 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2524186 T4811 oasc.SolrCore.<init> [nodes_used_collection_shard1_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/nodes_used_collection_shard1_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data/
   [junit4]   2> 2524186 T4811 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2524188 T4811 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data
   [junit4]   2> 2524189 T4811 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data/index/
   [junit4]   2> 2524189 T4811 oasc.SolrCore.initIndex WARN [nodes_used_collection_shard1_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/nodes_used_collection_shard1_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2524191 T4811 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data/index
   [junit4]   2> 2524205 T4766 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2524207 T4811 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@3b9d0ad lockFactory=org.apache.lucene.store.NativeFSLockFactory@f31d3d7),segFN=segments_1,generation=1}
   [junit4]   2> 2524208 T4811 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2524209 T4782 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2524221 T4796 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2524224 T4782 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2524225 T4782 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2524225 T4782 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2524222 T4811 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2524228 T4811 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2524229 T4811 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2524230 T4811 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2524231 T4811 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2524232 T4811 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2524232 T4811 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2524234 T4811 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2524235 T4811 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2524263 T4766 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2524267 T4811 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2524266 T4796 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2524263 T4782 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2524271 T4782 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2524271 T4782 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2524272 T4782 oasc.SolrCore.<init> [nodes_used_collection_shard1_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/nodes_used_collection_shard1_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data/
   [junit4]   2> 2524273 T4782 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2524275 T4811 oass.SolrIndexSearcher.<init> Opening Searcher@30da9a19 main
   [junit4]   2> 2524283 T4811 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2524284 T4811 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2524276 T4782 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data
   [junit4]   2> 2524288 T4782 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data/index/
   [junit4]   2> 2524289 T4782 oasc.SolrCore.initIndex WARN [nodes_used_collection_shard1_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/nodes_used_collection_shard1_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2524291 T4822 oasc.SolrCore.registerSearcher [nodes_used_collection_shard1_replica2] Registered new searcher Searcher@30da9a19 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2524291 T4766 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2524292 T4782 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data/index
   [junit4]   2> 2524297 T4796 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2524303 T4811 oasc.CoreContainer.registerCore registering core: nodes_used_collection_shard1_replica2
   [junit4]   2> 2524304 T4811 oasc.ZkController.register Register replica - core:nodes_used_collection_shard1_replica2 address:http://127.0.0.1:50812/i/v collection:nodes_used_collection shard:shard1
   [junit4]   2> 2524304 T4811 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2524307 T4796 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2524308 T4796 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2524309 T4796 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2524309 T4766 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2524311 T4766 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2524311 T4766 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2524314 T4782 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@140f9531 lockFactory=org.apache.lucene.store.NativeFSLockFactory@ea1dbcc),segFN=segments_1,generation=1}
   [junit4]   2> 2524314 T4782 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2524324 T4782 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2524325 T4782 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2524326 T4766 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2524327 T4766 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2524328 T4766 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2524328 T4766 oasc.SolrCore.<init> [nodes_used_collection_shard2_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/nodes_used_collection_shard2_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data/
   [junit4]   2> 2524329 T4766 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2524337 T4766 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data
   [junit4]   2> 2524337 T4766 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data/index/
   [junit4]   2> 2524341 T4766 oasc.SolrCore.initIndex WARN [nodes_used_collection_shard2_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/nodes_used_collection_shard2_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2524327 T4782 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2524343 T4782 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2524345 T4782 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2524345 T4782 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2524346 T4782 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2524347 T4782 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2524348 T4782 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2524330 T4796 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2524364 T4796 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2524365 T4796 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2524368 T4811 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection/leader_elect/shard1/election
   [junit4]   2> 2524369 T4782 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2524353 T4766 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data/index
   [junit4]   2> 2524369 T4796 oasc.SolrCore.<init> [nodes_used_collection_shard2_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/nodes_used_collection_shard2_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data/
   [junit4]   2> 2524381 T4796 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2524385 T4796 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data
   [junit4]   2> 2524386 T4796 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data/index/
   [junit4]   2> 2524386 T4796 oasc.SolrCore.initIndex WARN [nodes_used_collection_shard2_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/nodes_used_collection_shard2_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2524388 T4796 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data/index
   [junit4]   2> 2524393 T4766 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@1e7b8693 lockFactory=org.apache.lucene.store.NativeFSLockFactory@6e7b3aa4),segFN=segments_1,generation=1}
   [junit4]   2> 2524394 T4766 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2524394 T4782 oass.SolrIndexSearcher.<init> Opening Searcher@31e8855d main
   [junit4]   2> 2524397 T4782 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2524398 T4782 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2524405 T4766 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2524406 T4796 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@26e5fbcb lockFactory=org.apache.lucene.store.NativeFSLockFactory@39a02140),segFN=segments_1,generation=1}
   [junit4]   2> 2524407 T4796 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2524409 T4766 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2524412 T4823 oasc.SolrCore.registerSearcher [nodes_used_collection_shard1_replica1] Registered new searcher Searcher@31e8855d main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2524413 T4766 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2524414 T4766 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2524417 T4782 oasc.CoreContainer.registerCore registering core: nodes_used_collection_shard1_replica1
   [junit4]   2> 2524418 T4782 oasc.ZkController.register Register replica - core:nodes_used_collection_shard1_replica1 address:http://127.0.0.1:50806/i/v collection:nodes_used_collection shard:shard1
   [junit4]   2> 2524418 T4782 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2524417 T4766 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2524422 T4766 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2524423 T4766 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2524424 T4766 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2524425 T4766 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2524432 T4796 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2524433 T4796 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2524435 T4796 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2524436 T4796 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2524438 T4796 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2524438 T4796 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2524439 T4796 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2524440 T4796 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2524441 T4796 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2524451 T4796 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2524458 T4766 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2524465 T4766 oass.SolrIndexSearcher.<init> Opening Searcher@6287319d main
   [junit4]   2> 2524468 T4811 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 2524469 T4766 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2524472 T4766 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2524478 T4824 oasc.SolrCore.registerSearcher [nodes_used_collection_shard2_replica1] Registered new searcher Searcher@6287319d main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2524480 T4796 oass.SolrIndexSearcher.<init> Opening Searcher@3d54c573 main
   [junit4]   2> 2524481 T4766 oasc.CoreContainer.registerCore registering core: nodes_used_collection_shard2_replica1
   [junit4]   2> 2524482 T4766 oasc.ZkController.register Register replica - core:nodes_used_collection_shard2_replica1 address:http://127.0.0.1:50803/i/v collection:nodes_used_collection shard:shard2
   [junit4]   2> 2524483 T4796 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2524483 T4796 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2524483 T4766 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2524490 T4825 oasc.SolrCore.registerSearcher [nodes_used_collection_shard2_replica2] Registered new searcher Searcher@3d54c573 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2524495 T4796 oasc.CoreContainer.registerCore registering core: nodes_used_collection_shard2_replica2
   [junit4]   2> 2524496 T4796 oasc.ZkController.register Register replica - core:nodes_used_collection_shard2_replica2 address:http://127.0.0.1:50809/i/v collection:nodes_used_collection shard:shard2
   [junit4]   2> 2524497 T4796 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2524517 T4811 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 2524518 T4811 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2524519 T4811 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2/
   [junit4]   2> 2524520 T4811 oasu.PeerSync.sync PeerSync: core=nodes_used_collection_shard1_replica2 url=http://127.0.0.1:50812/i/v START replicas=[http://127.0.0.1:50806/i/v/nodes_used_collection_shard1_replica1/] nUpdates=100
   [junit4]   2> 2524527 T4811 oasu.PeerSync.sync PeerSync: core=nodes_used_collection_shard1_replica2 url=http://127.0.0.1:50812/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2524530 T4811 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2524535 T4811 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2524536 T4811 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2/ shard1
   [junit4]   2> 2524537 T4811 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection/leaders/shard1
   [junit4]   2> 2524540 T4766 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection/leader_elect/shard2/election
   [junit4]   2> 2524533 T4796 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection/leader_elect/shard2/election
   [junit4]   2> ASYNC  NEW_CORE C807 name=nodes_used_collection_shard1_replica1 org.apache.solr.core.SolrCore@4b3a2b0c url=http://127.0.0.1:50806/i/v/nodes_used_collection_shard1_replica1 node=127.0.0.1:50806_i%2Fv C807_STATE=coll:nodes_used_collection core:nodes_used_collection_shard1_replica1 props:{state=down, core=nodes_used_collection_shard1_replica1, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2524543 T4781 C807 P50806 oasc.SolrCore.execute [nodes_used_collection_shard1_replica1] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=2 
   [junit4]   2> 2524599 T4766 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 2524608 T4766 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 2524609 T4766 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2524610 T4766 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1/
   [junit4]   2> 2524610 T4766 oasu.PeerSync.sync PeerSync: core=nodes_used_collection_shard2_replica1 url=http://127.0.0.1:50803/i/v START replicas=[http://127.0.0.1:50809/i/v/nodes_used_collection_shard2_replica2/] nUpdates=100
   [junit4]   2> 2524611 T4766 oasu.PeerSync.sync PeerSync: core=nodes_used_collection_shard2_replica1 url=http://127.0.0.1:50803/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2524612 T4766 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2524612 T4766 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2524613 T4766 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1/ shard2
   [junit4]   2> 2524613 T4766 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection/leaders/shard2
   [junit4]   2> ASYNC  NEW_CORE C808 name=nodes_used_collection_shard2_replica2 org.apache.solr.core.SolrCore@14e7053b url=http://127.0.0.1:50809/i/v/nodes_used_collection_shard2_replica2 node=127.0.0.1:50809_i%2Fv C808_STATE=coll:nodes_used_collection core:nodes_used_collection_shard2_replica2 props:{state=down, core=nodes_used_collection_shard2_replica2, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2524632 T4794 C808 P50809 oasc.SolrCore.execute [nodes_used_collection_shard2_replica2] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=3 
   [junit4]   2> 2525873 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2525910 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2525911 T4761 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2525911 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2525912 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2525912 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2525913 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2525927 T4782 oasc.ZkController.register We are http://127.0.0.1:50806/i/v/nodes_used_collection_shard1_replica1/ and leader is http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2/
   [junit4]   2> 2525928 T4782 oasc.ZkController.register No LogReplay needed for core=nodes_used_collection_shard1_replica1 baseURL=http://127.0.0.1:50806/i/v
   [junit4]   2> 2525929 T4782 oasc.ZkController.checkRecovery Core needs to recover:nodes_used_collection_shard1_replica1
   [junit4]   2> 2525930 T4782 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2525933 T4828 C807 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=nodes_used_collection_shard1_replica1 recoveringAfterStartup=true
   [junit4]   2> 2525933 T4828 C807 P50806 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2525934 T4828 C807 P50806 oasc.ZkController.publish publishing core=nodes_used_collection_shard1_replica1 state=recovering
   [junit4]   2> 2525935 T4766 oasc.ZkController.register We are http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1/ and leader is http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1/
   [junit4]   2> 2525936 T4766 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 2525936 T4766 oasc.ZkController.publish publishing core=nodes_used_collection_shard2_replica1 state=active
   [junit4]   2> 2525937 T4811 oasc.ZkController.register We are http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2/ and leader is http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2/
   [junit4]   2> 2525938 T4811 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 2525938 T4811 oasc.ZkController.publish publishing core=nodes_used_collection_shard1_replica2 state=active
   [junit4]   2> 2525940 T4782 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2525943 T4828 C807 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2525954 T4782 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=2&shard=shard1&name=nodes_used_collection_shard1_replica1&action=CREATE&collection=nodes_used_collection&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=7580 
   [junit4]   2> 2525959 T4811 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2525960 T4766 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2525963 T4811 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=2&shard=shard1&name=nodes_used_collection_shard1_replica2&action=CREATE&collection=nodes_used_collection&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=7584 
   [junit4]   2> 2525965 T4766 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=2&shard=shard2&name=nodes_used_collection_shard2_replica1&action=CREATE&collection=nodes_used_collection&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=7600 
   [junit4]   2> 2525966 T4796 oasc.ZkController.register We are http://127.0.0.1:50809/i/v/nodes_used_collection_shard2_replica2/ and leader is http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1/
   [junit4]   2> 2525966 T4796 oasc.ZkController.register No LogReplay needed for core=nodes_used_collection_shard2_replica2 baseURL=http://127.0.0.1:50809/i/v
   [junit4]   2> 2525967 T4796 oasc.ZkController.checkRecovery Core needs to recover:nodes_used_collection_shard2_replica2
   [junit4]   2> 2525967 T4796 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2525970 T4829 C808 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=nodes_used_collection_shard2_replica2 recoveringAfterStartup=true
   [junit4]   2> 2525972 T4829 C808 P50809 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2525972 T4829 C808 P50809 oasc.ZkController.publish publishing core=nodes_used_collection_shard2_replica2 state=recovering
   [junit4]   2> 2525975 T4796 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2525977 T4829 C808 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2525992 T4796 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=2&shard=shard2&name=nodes_used_collection_shard2_replica2&action=CREATE&collection=nodes_used_collection&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=7609 
   [junit4]   2> 2525997 T4758 oasc.OverseerCollectionProcessor.createCollection Finished create command on all shards for collection: nodes_used_collection
   [junit4]   2> 2526012 T4810 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node2, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2526024 T4769 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node4, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2526029 T4774 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work/qnr-0000000000 state: SyncConnected type NodeDataChanged
   [junit4]   2> 2526033 T4756 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work state: SyncConnected type NodeChildrenChanged
   [junit4]   2> 2526036 T4765 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/collections params={numShards=2&name=nodes_used_collection&replicationFactor=2&action=CREATE&wt=javabin&version=2} status=0 QTime=7701 
   [junit4]   2> 2526038 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:
   [junit4]   2> 2526040 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2526041 T4758 oasc.OverseerCollectionProcessor.run Overseer Collection Processor: Message id:/overseer/collection-queue-work/qn-0000000000 complete, response:{success={null={responseHeader={status=0,QTime=7580},core=nodes_used_collection_shard1_replica1},null={responseHeader={status=0,QTime=7584},core=nodes_used_collection_shard1_replica2},null={responseHeader={status=0,QTime=7600},core=nodes_used_collection_shard2_replica1},null={responseHeader={status=0,QTime=7609},core=nodes_used_collection_shard2_replica2}}}
   [junit4]   2> 2526048 T4831 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@13133821 name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2526049 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2526052 T4734 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 2526060 T4734 oascsi.HttpClientUtil.createClient Creating new http client, config:
   [junit4]   2> 2526063 T4734 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 2526070 T4833 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@7a909ade name:ZooKeeperConnection Watcher:127.0.0.1:50796/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 2526071 T4734 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 2526074 T4734 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 2526113 T4812 oasha.CollectionsHandler.handleCreateAction Creating Collection : numShards=5&maxShardsPerNode=6&name=awholynewcollection_0&replicationFactor=5&action=CREATE&wt=javabin&version=2
   [junit4]   2> 2526122 T4756 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work state: SyncConnected type NodeChildrenChanged
   [junit4]   2> 2526127 T4758 oasc.OverseerCollectionProcessor.run Overseer Collection Processor: Get the message id:/overseer/collection-queue-work/qn-0000000002 message:{
   [junit4]   2> 	  "operation":"createcollection",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "maxShardsPerNode":"6",
   [junit4]   2> 	  "createNodeSet":null,
   [junit4]   2> 	  "name":"awholynewcollection_0",
   [junit4]   2> 	  "replicationFactor":"5"}
   [junit4]   2> 2526127 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard1_replica1 as part of slice shard1 of collection awholynewcollection_0 on 127.0.0.1:50809_i%2Fv
   [junit4]   2> 2526128 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard1_replica2 as part of slice shard1 of collection awholynewcollection_0 on 127.0.0.1:50799_i%2Fv
   [junit4]   2> 2526129 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard1_replica3 as part of slice shard1 of collection awholynewcollection_0 on 127.0.0.1:50806_i%2Fv
   [junit4]   2> 2526131 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard1_replica4 as part of slice shard1 of collection awholynewcollection_0 on 127.0.0.1:50803_i%2Fv
   [junit4]   2> 2526134 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard1_replica5 as part of slice shard1 of collection awholynewcollection_0 on 127.0.0.1:50812_i%2Fv
   [junit4]   2> 2526135 T4795 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica1 state=down
   [junit4]   2> 2526136 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard2_replica1 as part of slice shard2 of collection awholynewcollection_0 on 127.0.0.1:50809_i%2Fv
   [junit4]   2> 2526141 T4770 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica4 state=down
   [junit4]   2> 2526139 T4807 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica5 state=down
   [junit4]   2> 2526139 T4780 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica3 state=down
   [junit4]   2> 2526143 T4752 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica2 state=down
   [junit4]   2> 2526148 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard2_replica2 as part of slice shard2 of collection awholynewcollection_0 on 127.0.0.1:50799_i%2Fv
   [junit4]   2> 2526152 T4794 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica1 state=down
   [junit4]   2> 2526157 T4795 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526158 T4807 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526159 T4752 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526160 T4747 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica2 state=down
   [junit4]   2> 2526162 T4770 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526163 T4780 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526153 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard2_replica3 as part of slice shard2 of collection awholynewcollection_0 on 127.0.0.1:50806_i%2Fv
   [junit4]   2> 2526166 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard2_replica4 as part of slice shard2 of collection awholynewcollection_0 on 127.0.0.1:50803_i%2Fv
   [junit4]   2> 2526169 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard2_replica5 as part of slice shard2 of collection awholynewcollection_0 on 127.0.0.1:50812_i%2Fv
   [junit4]   2> 2526174 T4794 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526174 T4747 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526174 T4783 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica3 state=down
   [junit4]   2> 2526180 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard3_replica1 as part of slice shard3 of collection awholynewcollection_0 on 127.0.0.1:50809_i%2Fv
   [junit4]   2> 2526181 T4808 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica5 state=down
   [junit4]   2> 2526173 T4767 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica4 state=down
   [junit4]   2> 2526185 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard3_replica2 as part of slice shard3 of collection awholynewcollection_0 on 127.0.0.1:50799_i%2Fv
   [junit4]   2> 2526192 T4767 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526191 T4808 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526190 T4783 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526189 T4797 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica1 state=down
   [junit4]   2> 2526195 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard3_replica3 as part of slice shard3 of collection awholynewcollection_0 on 127.0.0.1:50806_i%2Fv
   [junit4]   2> 2526200 T4748 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica2 state=down
   [junit4]   2> 2526203 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard3_replica4 as part of slice shard3 of collection awholynewcollection_0 on 127.0.0.1:50803_i%2Fv
   [junit4]   2> 2526207 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard3_replica5 as part of slice shard3 of collection awholynewcollection_0 on 127.0.0.1:50812_i%2Fv
   [junit4]   2> 2526213 T4748 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526215 T4766 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica4 state=down
   [junit4]   2> 2526216 T4797 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526212 T4784 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica3 state=down
   [junit4]   2> 2526218 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard4_replica1 as part of slice shard4 of collection awholynewcollection_0 on 127.0.0.1:50809_i%2Fv
   [junit4]   2> 2526222 T4811 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica5 state=down
   [junit4]   2> 2526225 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard4_replica2 as part of slice shard4 of collection awholynewcollection_0 on 127.0.0.1:50799_i%2Fv
   [junit4]   2> 2526231 T4811 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526231 T4798 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=down
   [junit4]   2> 2526228 T4766 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526228 T4784 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526239 T4749 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica2 state=down
   [junit4]   2> 2526239 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard4_replica3 as part of slice shard4 of collection awholynewcollection_0 on 127.0.0.1:50806_i%2Fv
   [junit4]   2> 2526242 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard4_replica4 as part of slice shard4 of collection awholynewcollection_0 on 127.0.0.1:50803_i%2Fv
   [junit4]   2> 2526246 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard4_replica5 as part of slice shard4 of collection awholynewcollection_0 on 127.0.0.1:50812_i%2Fv
   [junit4]   2> 2526254 T4779 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=down
   [junit4]   2> 2526252 T4768 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica4 state=down
   [junit4]   2> 2526250 T4798 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526263 T4779 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526264 T4749 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526264 T4809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=down
   [junit4]   2> 2526267 T4768 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526270 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard5_replica1 as part of slice shard5 of collection awholynewcollection_0 on 127.0.0.1:50809_i%2Fv
   [junit4]   2> 2526270 T4809 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526273 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard5_replica2 as part of slice shard5 of collection awholynewcollection_0 on 127.0.0.1:50799_i%2Fv
   [junit4]   2> 2526279 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard5_replica3 as part of slice shard5 of collection awholynewcollection_0 on 127.0.0.1:50806_i%2Fv
   [junit4]   2> 2526280 T4796 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica1 state=down
   [junit4]   2> 2526284 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard5_replica4 as part of slice shard5 of collection awholynewcollection_0 on 127.0.0.1:50803_i%2Fv
   [junit4]   2> 2526295 T4750 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica2 state=down
   [junit4]   2> 2526295 T4782 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica3 state=down
   [junit4]   2> 2526290 T4796 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526301 T4758 oasc.OverseerCollectionProcessor.createCollection Creating shard awholynewcollection_0_shard5_replica5 as part of slice shard5 of collection awholynewcollection_0 on 127.0.0.1:50812_i%2Fv
   [junit4]   2> 2526303 T4782 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526303 T4765 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica4 state=down
   [junit4]   2> 2526309 T4750 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526311 T4855 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica5 state=down
   [junit4]   2> 2526312 T4765 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2526314 T4855 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 2527467 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2527471 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard1_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2527480 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard1_replica2",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2527488 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2527497 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica2",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2527506 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527507 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2527507 T4757 oasc.Overseer$ClusterStateUpdater.createCollection Create collection awholynewcollection_0 with numShards 5
   [junit4]   2> 2527516 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527516 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2527530 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527531 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2527542 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527543 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2527559 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527560 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2527573 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527573 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2527582 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527583 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2527591 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527591 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2527612 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527612 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node9",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2527622 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527623 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node10",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2527635 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527635 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node11",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2527644 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527644 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node12",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2527659 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527660 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node13",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2527677 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527678 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node14",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2527687 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527688 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node15",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2527696 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527696 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2527782 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527783 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node17",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2527792 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527793 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2527803 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527804 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node19",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2527813 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527813 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2527822 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527823 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node21",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2527831 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527832 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node22",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2527840 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527840 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node23",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2527849 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527849 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node24",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2527857 T4757 oasc.Overseer$ClusterStateUpdater.updateState WARN Could not find collection node for awholynewcollection_0, skipping publish state
   [junit4]   2> 2527858 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node25",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2527866 T4831 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2527866 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2527866 T4761 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2527868 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2527868 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2527869 T4833 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2527870 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2527870 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2528016 T4810 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node2, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2528017 T4810 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node2&state=recovering&nodeName=127.0.0.1:50806_i%252Fv&action=PREPRECOVERY&checkLive=true&core=nodes_used_collection_shard1_replica2&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2005 
   [junit4]   2> 2528027 T4769 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node4, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2528028 T4769 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node4&state=recovering&nodeName=127.0.0.1:50809_i%252Fv&action=PREPRECOVERY&checkLive=true&core=nodes_used_collection_shard2_replica1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2004 
   [junit4]   2> 2528161 T4795 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard1_replica1
   [junit4]   2> 2528162 T4752 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard1_replica2
   [junit4]   2> 2528163 T4795 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard1_replica1/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard1_replica1/core.properties (No such file or directory)
   [junit4]   2> 2528163 T4795 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard1_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard1_replica1
   [junit4]   2> 2528164 T4795 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528163 T4752 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard1_replica2/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard1_replica2/core.properties (No such file or directory)
   [junit4]   2> 2528165 T4752 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard1_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard1_replica2
   [junit4]   2> 2528166 T4752 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528164 T4807 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard1_replica5
   [junit4]   2> 2528165 T4780 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard1_replica3
   [junit4]   2> 2528167 T4807 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard1_replica5/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard1_replica5/core.properties (No such file or directory)
   [junit4]   2> 2528168 T4807 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard1_replica5' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard1_replica5
   [junit4]   2> 2528168 T4807 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528169 T4795 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:awholynewcollection_0
   [junit4]   2> 2528170 T4795 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2528165 T4770 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard1_replica4
   [junit4]   2> 2528168 T4780 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard1_replica3/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard1_replica3/core.properties (No such file or directory)
   [junit4]   2> 2528172 T4780 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard1_replica3' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard1_replica3
   [junit4]   2> 2528173 T4780 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528173 T4770 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard1_replica4/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard1_replica4/core.properties (No such file or directory)
   [junit4]   2> 2528174 T4770 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard1_replica4' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard1_replica4
   [junit4]   2> 2528175 T4770 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528174 T4752 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:awholynewcollection_0
   [junit4]   2> 2528176 T4752 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2528176 T4807 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:awholynewcollection_0
   [junit4]   2> 2528177 T4780 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:awholynewcollection_0
   [junit4]   2> 2528177 T4807 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2528178 T4747 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard2_replica2
   [junit4]   2> 2528181 T4747 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard2_replica2/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard2_replica2/core.properties (No such file or directory)
   [junit4]   2> 2528181 T4747 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard2_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard2_replica2
   [junit4]   2> 2528182 T4747 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528178 T4794 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard2_replica1
   [junit4]   2> 2528178 T4780 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2528183 T4794 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard2_replica1/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard2_replica1/core.properties (No such file or directory)
   [junit4]   2> 2528184 T4794 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard2_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard2_replica1
   [junit4]   2> 2528184 T4794 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528182 T4795 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2528186 T4795 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0
   [junit4]   2> 2528181 T4807 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2528188 T4807 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0
   [junit4]   2> 2528179 T4770 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:awholynewcollection_0
   [junit4]   2> 2528189 T4747 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:awholynewcollection_0
   [junit4]   2> 2528189 T4794 oasc.ZkController.createCollectionZkNode Creating collection in ZooKeeper:awholynewcollection_0
   [junit4]   2> 2528190 T4794 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2528189 T4752 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2528191 T4752 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0
   [junit4]   2> 2528189 T4770 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2528189 T4747 oasc.ZkController.getConfName Looking for collection configName
   [junit4]   2> 2528194 T4780 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2528195 T4780 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0
   [junit4]   2> 2528197 T4767 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard2_replica4
   [junit4]   2> 2528198 T4767 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard2_replica4/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard2_replica4/core.properties (No such file or directory)
   [junit4]   2> 2528199 T4767 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard2_replica4' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard2_replica4
   [junit4]   2> 2528199 T4767 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528199 T4808 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard2_replica5
   [junit4]   2> 2528201 T4747 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2528201 T4747 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0
   [junit4]   2> 2528202 T4770 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2528202 T4770 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0
   [junit4]   2> 2528202 T4783 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard2_replica3
   [junit4]   2> 2528204 T4783 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard2_replica3/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard2_replica3/core.properties (No such file or directory)
   [junit4]   2> 2528205 T4783 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard2_replica3' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard2_replica3
   [junit4]   2> 2528206 T4783 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528201 T4808 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard2_replica5/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard2_replica5/core.properties (No such file or directory)
   [junit4]   2> 2528207 T4808 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard2_replica5' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard2_replica5
   [junit4]   2> 2528207 T4808 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528202 T4794 oasc.ZkController.getConfName Only one config set found in zk - using it:conf1
   [junit4]   2> 2528208 T4794 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0
   [junit4]   2> 2528211 T4795 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528211 T4807 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528211 T4767 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528215 T4767 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528211 T4783 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528217 T4783 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528211 T4752 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528220 T4797 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard3_replica1
   [junit4]   2> 2528222 T4797 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard3_replica1/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard3_replica1/core.properties (No such file or directory)
   [junit4]   2> 2528223 T4797 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard3_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard3_replica1
   [junit4]   2> 2528223 T4797 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528225 T4752 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard1_replica2/'
   [junit4]   2> 2528225 T4747 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528217 T4780 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528217 T4748 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard3_replica2
   [junit4]   2> 2528229 T4748 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard3_replica2/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard3_replica2/core.properties (No such file or directory)
   [junit4]   2> 2528229 T4748 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard3_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard3_replica2
   [junit4]   2> 2528230 T4748 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528235 T4811 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard3_replica5
   [junit4]   2> 2528216 T4808 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528237 T4808 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528235 T4795 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard1_replica1/'
   [junit4]   2> 2528244 T4811 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard3_replica5/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard3_replica5/core.properties (No such file or directory)
   [junit4]   2> 2528245 T4811 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard3_replica5' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard3_replica5
   [junit4]   2> 2528246 T4811 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528225 T4807 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard1_replica5/'
   [junit4]   2> 2528248 T4766 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard3_replica4
   [junit4]   2> 2528249 T4766 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard3_replica4/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard3_replica4/core.properties (No such file or directory)
   [junit4]   2> 2528250 T4766 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard3_replica4' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard3_replica4
   [junit4]   2> 2528250 T4766 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528251 T4784 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard3_replica3
   [junit4]   2> 2528252 T4784 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard3_replica3/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard3_replica3/core.properties (No such file or directory)
   [junit4]   2> 2528253 T4784 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard3_replica3' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard3_replica3
   [junit4]   2> 2528253 T4784 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528263 T4798 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard4_replica1
   [junit4]   2> 2528265 T4798 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard4_replica1/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard4_replica1/core.properties (No such file or directory)
   [junit4]   2> 2528267 T4798 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard4_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard4_replica1
   [junit4]   2> 2528268 T4798 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528267 T4749 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard4_replica2
   [junit4]   2> 2528267 T4779 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard4_replica3
   [junit4]   2> 2528270 T4749 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard4_replica2/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard4_replica2/core.properties (No such file or directory)
   [junit4]   2> 2528271 T4749 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard4_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard4_replica2
   [junit4]   2> 2528271 T4749 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528272 T4768 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard4_replica4
   [junit4]   2> 2528272 T4779 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard4_replica3/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard4_replica3/core.properties (No such file or directory)
   [junit4]   2> 2528274 T4779 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard4_replica3' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard4_replica3
   [junit4]   2> 2528274 T4779 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528274 T4809 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard4_replica5
   [junit4]   2> 2528277 T4809 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard4_replica5/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard4_replica5/core.properties (No such file or directory)
   [junit4]   2> 2528279 T4809 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard4_replica5' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard4_replica5
   [junit4]   2> 2528285 T4809 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528289 T4797 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528290 T4797 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528274 T4768 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard4_replica4/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard4_replica4/core.properties (No such file or directory)
   [junit4]   2> 2528291 T4768 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard4_replica4' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard4_replica4
   [junit4]   2> 2528292 T4768 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528291 T4794 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528293 T4767 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard2_replica4/'
   [junit4]   2> 2528289 T4783 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard2_replica3/'
   [junit4]   2> 2528303 T4796 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard5_replica1
   [junit4]   2> 2528305 T4796 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard5_replica1/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard5_replica1/core.properties (No such file or directory)
   [junit4]   2> 2528306 T4796 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard5_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard5_replica1
   [junit4]   2> 2528307 T4796 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528308 T4782 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard5_replica3
   [junit4]   2> 2528308 T4784 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528309 T4784 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528308 T4811 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528310 T4811 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528309 T4782 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard5_replica3/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard5_replica3/core.properties (No such file or directory)
   [junit4]   2> 2528312 T4782 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard5_replica3' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard5_replica3
   [junit4]   2> 2528312 T4782 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528313 T4779 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528313 T4779 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528312 T4770 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528316 T4765 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard5_replica4
   [junit4]   2> 2528317 T4765 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard5_replica4/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard5_replica4/core.properties (No such file or directory)
   [junit4]   2> 2528318 T4765 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard5_replica4' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard5_replica4
   [junit4]   2> 2528318 T4765 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528314 T4798 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528319 T4798 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528313 T4766 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528320 T4766 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528313 T4750 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard5_replica2
   [junit4]   2> 2528323 T4768 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528323 T4768 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528317 T4780 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard1_replica3/'
   [junit4]   2> 2528317 T4855 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for awholynewcollection_0_shard5_replica5
   [junit4]   2> 2528325 T4808 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard2_replica5/'
   [junit4]   2> 2528324 T4796 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528327 T4796 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528326 T4782 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528329 T4782 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528332 T4855 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard5_replica5/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard5_replica5/core.properties (No such file or directory)
   [junit4]   2> 2528332 T4855 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard5_replica5' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard5_replica5
   [junit4]   2> 2528333 T4855 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528333 T4750 oasc.CorePropertiesLocator.writePropertiesFile ERROR Couldn't persist core properties to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard5_replica2/core.properties: java.io.FileNotFoundException: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard5_replica2/core.properties (No such file or directory)
   [junit4]   2> 2528334 T4750 oasc.CoreContainer.create Creating SolrCore 'awholynewcollection_0_shard5_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard5_replica2
   [junit4]   2> 2528334 T4750 oasc.ZkController.createCollectionZkNode Check for collection zkNode:awholynewcollection_0
   [junit4]   2> 2528338 T4765 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528338 T4765 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528340 T4784 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard3_replica3/'
   [junit4]   2> 2528342 T4797 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard3_replica1/'
   [junit4]   2> 2528348 T4811 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard3_replica5/'
   [junit4]   2> 2528328 T4809 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528352 T4809 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528355 T4855 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528356 T4855 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528357 T4794 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard2_replica1/'
   [junit4]   2> 2528369 T4779 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard4_replica3/'
   [junit4]   2> 2528369 T4766 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard3_replica4/'
   [junit4]   2> 2528369 T4796 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard5_replica1/'
   [junit4]   2> 2528369 T4798 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard4_replica1/'
   [junit4]   2> 2528373 T4768 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard4_replica4/'
   [junit4]   2> 2528373 T4770 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard1_replica4/'
   [junit4]   2> 2528379 T4782 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard5_replica3/'
   [junit4]   2> 2528382 T4855 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard5_replica5/'
   [junit4]   2> 2528384 T4765 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard5_replica4/'
   [junit4]   2> 2528386 T4809 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard4_replica5/'
   [junit4]   2> 2528386 T4748 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528388 T4748 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528390 T4749 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528390 T4749 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528391 T4747 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard2_replica2/'
   [junit4]   2> 2528391 T4750 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 2528401 T4750 oasc.ZkController.readConfigName Load collection config from:/collections/awholynewcollection_0
   [junit4]   2> 2528417 T4748 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard3_replica2/'
   [junit4]   2> 2528423 T4749 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard4_replica2/'
   [junit4]   2> 2528432 T4750 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard5_replica2/'
   [junit4]   2> 2528931 T4795 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529030 T4808 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529077 T4767 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529205 T4811 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529252 T4794 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529285 T4780 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529299 T4784 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529319 T4752 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529571 T4809 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529578 T4807 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529762 T4783 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529867 T4747 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529895 T4796 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529921 T4779 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2529927 T4795 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2529968 T4795 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> ASYNC  NEW_CORE C809 name=nodes_used_collection_shard1_replica1 org.apache.solr.core.SolrCore@4b3a2b0c url=http://127.0.0.1:50806/i/v/nodes_used_collection_shard1_replica1 node=127.0.0.1:50806_i%2Fv C809_STATE=coll:nodes_used_collection core:nodes_used_collection_shard1_replica1 props:{state=recovering, core=nodes_used_collection_shard1_replica1, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2530019 T4828 C809 P50806 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2/ core=nodes_used_collection_shard1_replica1 - recoveringAfterStartup=true
   [junit4]   2> 2530020 T4828 C809 P50806 oasu.PeerSync.sync PeerSync: core=nodes_used_collection_shard1_replica1 url=http://127.0.0.1:50806/i/v START replicas=[http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2/] nUpdates=100
   [junit4]   2> 2530022 T4828 C809 P50806 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2530022 T4828 C809 P50806 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=nodes_used_collection_shard1_replica1
   [junit4]   2> 2530022 T4828 C809 P50806 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=nodes_used_collection_shard1_replica1
   [junit4]   2> 2530023 T4828 C809 P50806 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=nodes_used_collection_shard1_replica1
   [junit4]   2> 2530023 T4828 C809 P50806 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2530024 T4828 C809 P50806 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2/. core=nodes_used_collection_shard1_replica1
   [junit4]   2> 2530024 T4828 C809 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> ASYNC  NEW_CORE C810 name=nodes_used_collection_shard2_replica2 org.apache.solr.core.SolrCore@14e7053b url=http://127.0.0.1:50809/i/v/nodes_used_collection_shard2_replica2 node=127.0.0.1:50809_i%2Fv C810_STATE=coll:nodes_used_collection core:nodes_used_collection_shard2_replica2 props:{state=recovering, core=nodes_used_collection_shard2_replica2, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2530031 T4829 C810 P50809 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1/ core=nodes_used_collection_shard2_replica2 - recoveringAfterStartup=true
   [junit4]   2> 2530052 T4829 C810 P50809 oasu.PeerSync.sync PeerSync: core=nodes_used_collection_shard2_replica2 url=http://127.0.0.1:50809/i/v START replicas=[http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1/] nUpdates=100
   [junit4]   2> 2530046 T4795 oass.IndexSchema.readSchema [awholynewcollection_0_shard1_replica1] Schema name=test
   [junit4]   2> 2530063 T4829 C810 P50809 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2530063 T4829 C810 P50809 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=nodes_used_collection_shard2_replica2
   [junit4]   2> 2530063 T4829 C810 P50809 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=nodes_used_collection_shard2_replica2
   [junit4]   2> 2530064 T4829 C810 P50809 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=nodes_used_collection_shard2_replica2
   [junit4]   2> 2530064 T4829 C810 P50809 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2530064 T4829 C810 P50809 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1/. core=nodes_used_collection_shard2_replica2
   [junit4]   2> 2530065 T4829 C810 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> ASYNC  NEW_CORE C811 name=nodes_used_collection_shard1_replica2 org.apache.solr.core.SolrCore@2c000624 url=http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2 node=127.0.0.1:50812_i%2Fv C811_STATE=coll:nodes_used_collection core:nodes_used_collection_shard1_replica2 props:{state=active, core=nodes_used_collection_shard1_replica2, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v, leader=true}
   [junit4]   2> 2530065 T4859 C811 P50812 oasc.SolrCore.execute [nodes_used_collection_shard1_replica2] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2530082 T4855 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2530083 T4782 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> ASYNC  NEW_CORE C812 name=nodes_used_collection_shard2_replica1 org.apache.solr.core.SolrCore@22cda955 url=http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1 node=127.0.0.1:50803_i%2Fv C812_STATE=coll:nodes_used_collection core:nodes_used_collection_shard2_replica1 props:{state=active, core=nodes_used_collection_shard2_replica1, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v, leader=true}
   [junit4]   2> 2530120 T4858 C812 P50803 oasc.SolrCore.execute [nodes_used_collection_shard2_replica1] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> ASYNC  NEW_CORE C813 name=nodes_used_collection_shard1_replica2 org.apache.solr.core.SolrCore@2c000624 url=http://127.0.0.1:50812/i/v/nodes_used_collection_shard1_replica2 node=127.0.0.1:50812_i%2Fv C813_STATE=coll:nodes_used_collection core:nodes_used_collection_shard1_replica2 props:{state=active, core=nodes_used_collection_shard1_replica2, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v, leader=true}
   [junit4]   2> 2530266 T4810 C813 P50812 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> ASYNC  NEW_CORE C814 name=nodes_used_collection_shard2_replica1 org.apache.solr.core.SolrCore@22cda955 url=http://127.0.0.1:50803/i/v/nodes_used_collection_shard2_replica1 node=127.0.0.1:50803_i%2Fv C814_STATE=coll:nodes_used_collection core:nodes_used_collection_shard2_replica1 props:{state=active, core=nodes_used_collection_shard2_replica1, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v, leader=true}
   [junit4]   2> 2530277 T4769 C814 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2530296 T4765 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2530318 T4769 C814 P50803 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@1e7b8693 lockFactory=org.apache.lucene.store.NativeFSLockFactory@6e7b3aa4),segFN=segments_1,generation=1}
   [junit4]   2> 2530321 T4769 C814 P50803 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2530331 T4769 C814 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2530334 T4769 C814 P50803 oass.SolrIndexSearcher.<init> Opening Searcher@54dd5350 realtime
   [junit4]   2> 2530335 T4769 C814 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2530335 T4769 C814 P50803 oasup.LogUpdateProcessor.finish [nodes_used_collection_shard2_replica1] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 58
   [junit4]   2> ASYNC  NEW_CORE C815 name=nodes_used_collection_shard2_replica2 org.apache.solr.core.SolrCore@14e7053b url=http://127.0.0.1:50809/i/v/nodes_used_collection_shard2_replica2 node=127.0.0.1:50809_i%2Fv C815_STATE=coll:nodes_used_collection core:nodes_used_collection_shard2_replica2 props:{state=recovering, core=nodes_used_collection_shard2_replica2, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2530337 T4829 C815 P50809 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2530338 T4829 C815 P50809 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2530358 T4769 C814 P50803 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2530359 T4769 C814 P50803 oasc.SolrCore.execute [nodes_used_collection_shard2_replica1] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=1 
   [junit4]   2> 2530370 T4829 C815 P50809 oasc.RecoveryStrategy.replay No replay needed. core=nodes_used_collection_shard2_replica2
   [junit4]   2> 2530371 T4829 C815 P50809 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=nodes_used_collection_shard2_replica2
   [junit4]   2> 2530371 T4829 C815 P50809 oasc.ZkController.publish publishing core=nodes_used_collection_shard2_replica2 state=active
   [junit4]   2> 2530376 T4829 C815 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=nodes_used_collection_shard2_replica2
   [junit4]   2> 2530383 T4807 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2530411 T4807 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2530472 T4807 oass.IndexSchema.readSchema [awholynewcollection_0_shard1_replica5] Schema name=test
   [junit4]   2> 2530482 T4810 C813 P50812 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@3b9d0ad lockFactory=org.apache.lucene.store.NativeFSLockFactory@f31d3d7),segFN=segments_1,generation=1}
   [junit4]   2> 2530519 T4810 C813 P50812 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2530522 T4810 C813 P50812 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2530523 T4810 C813 P50812 oass.SolrIndexSearcher.<init> Opening Searcher@33d7e6ee realtime
   [junit4]   2> 2530524 T4810 C813 P50812 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2530524 T4810 C813 P50812 oasup.LogUpdateProcessor.finish [nodes_used_collection_shard1_replica2] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 258
   [junit4]   2> ASYNC  NEW_CORE C816 name=nodes_used_collection_shard1_replica1 org.apache.solr.core.SolrCore@4b3a2b0c url=http://127.0.0.1:50806/i/v/nodes_used_collection_shard1_replica1 node=127.0.0.1:50806_i%2Fv C816_STATE=coll:nodes_used_collection core:nodes_used_collection_shard1_replica1 props:{state=recovering, core=nodes_used_collection_shard1_replica1, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2530526 T4828 C816 P50806 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2530527 T4828 C816 P50806 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2530547 T4859 C813 P50812 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2530548 T4859 C813 P50812 oasc.SolrCore.execute [nodes_used_collection_shard1_replica2] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=3 
   [junit4]   2> 2530556 T4828 C816 P50806 oasc.RecoveryStrategy.replay No replay needed. core=nodes_used_collection_shard1_replica1
   [junit4]   2> 2530556 T4828 C816 P50806 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=nodes_used_collection_shard1_replica1
   [junit4]   2> 2530559 T4828 C816 P50806 oasc.ZkController.publish publishing core=nodes_used_collection_shard1_replica1 state=active
   [junit4]   2> 2530563 T4828 C816 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=nodes_used_collection_shard1_replica1
   [junit4]   2> 2530584 T4766 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2530665 T4752 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2530689 T4752 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2530753 T4752 oass.IndexSchema.readSchema [awholynewcollection_0_shard1_replica2] Schema name=test
   [junit4]   2> 2530836 T4748 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2530837 T4808 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2530867 T4808 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2530880 T4794 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2530919 T4767 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2530922 T4794 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2530946 T4767 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2530966 T4750 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2530994 T4768 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2531018 T4798 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2531024 T4794 oass.IndexSchema.readSchema [awholynewcollection_0_shard2_replica1] Schema name=test
   [junit4]   2> 2531219 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2531231 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica2",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2531259 T4749 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2531263 T4784 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2531280 T4784 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2531334 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard1_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2531345 T4767 oass.IndexSchema.readSchema [awholynewcollection_0_shard2_replica4] Schema name=test
   [junit4]   2> 2531346 T4796 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2531354 T4796 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2531399 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2531400 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2531400 T4831 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2531400 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2531400 T4833 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2531401 T4761 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2531400 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2531417 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2531449 T4808 oass.IndexSchema.readSchema [awholynewcollection_0_shard2_replica5] Schema name=test
   [junit4]   2> 2531459 T4796 oass.IndexSchema.readSchema [awholynewcollection_0_shard5_replica1] Schema name=test
   [junit4]   2> 2531534 T4747 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2531541 T4747 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2531624 T4765 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2531629 T4784 oass.IndexSchema.readSchema [awholynewcollection_0_shard3_replica3] Schema name=test
   [junit4]   2> 2531632 T4765 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2531636 T4797 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2531650 T4747 oass.IndexSchema.readSchema [awholynewcollection_0_shard2_replica2] Schema name=test
   [junit4]   2> 2531666 T4765 oass.IndexSchema.readSchema [awholynewcollection_0_shard5_replica4] Schema name=test
   [junit4]   2> 2531740 T4766 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2531745 T4783 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2531770 T4783 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2531804 T4809 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2531821 T4809 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2531838 T4809 oass.IndexSchema.readSchema [awholynewcollection_0_shard4_replica5] Schema name=test
   [junit4]   2> 2531841 T4766 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2531845 T4783 oass.IndexSchema.readSchema [awholynewcollection_0_shard2_replica3] Schema name=test
   [junit4]   2> 2532003 T4780 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2532010 T4780 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2532052 T4780 oass.IndexSchema.readSchema [awholynewcollection_0_shard1_replica3] Schema name=test
   [junit4]   2> 2532079 T4770 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 2532090 T4811 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2532101 T4766 oass.IndexSchema.readSchema [awholynewcollection_0_shard3_replica4] Schema name=test
   [junit4]   2> 2532108 T4782 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2532102 T4811 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2532131 T4782 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2532147 T4782 oass.IndexSchema.readSchema [awholynewcollection_0_shard5_replica3] Schema name=test
   [junit4]   2> 2532283 T4811 oass.IndexSchema.readSchema [awholynewcollection_0_shard3_replica5] Schema name=test
   [junit4]   2> 2532460 T4855 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2532484 T4855 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2532755 T4855 oass.IndexSchema.readSchema [awholynewcollection_0_shard5_replica5] Schema name=test
   [junit4]   2> 2532964 T4750 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2532972 T4750 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2532988 T4768 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2532993 T4768 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2533053 T4798 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2533085 T4798 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2533122 T4768 oass.IndexSchema.readSchema [awholynewcollection_0_shard4_replica4] Schema name=test
   [junit4]   2> 2533159 T4779 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2533168 T4779 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2533178 T4748 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2533187 T4779 oass.IndexSchema.readSchema [awholynewcollection_0_shard4_replica3] Schema name=test
   [junit4]   2> 2533202 T4748 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2533233 T4749 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2533249 T4749 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2533264 T4749 oass.IndexSchema.readSchema [awholynewcollection_0_shard4_replica2] Schema name=test
   [junit4]   2> 2533302 T4748 oass.IndexSchema.readSchema [awholynewcollection_0_shard3_replica2] Schema name=test
   [junit4]   2> 2533539 T4750 oass.IndexSchema.readSchema [awholynewcollection_0_shard5_replica2] Schema name=test
   [junit4]   2> 2533558 T4798 oass.IndexSchema.readSchema [awholynewcollection_0_shard4_replica1] Schema name=test
   [junit4]   2> 2533723 T4770 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2533739 T4770 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2533859 T4770 oass.IndexSchema.readSchema [awholynewcollection_0_shard1_replica4] Schema name=test
   [junit4]   2> 2534260 T4797 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 2534274 T4797 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 2534746 T4797 oass.IndexSchema.readSchema [awholynewcollection_0_shard3_replica1] Schema name=test
   [junit4]   2> 2550748 T4807 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2550770 T4807 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2550787 T4807 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2551211 T4807 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2551280 T4767 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2551300 T4767 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2551339 T4807 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2551469 T4807 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2551535 T4807 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2551536 T4807 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2551537 T4807 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2551540 T4807 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2551541 T4807 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2551542 T4807 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2551543 T4807 oasc.SolrCore.<init> [awholynewcollection_0_shard1_replica5] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard1_replica5/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data/
   [junit4]   2> 2551544 T4807 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2551548 T4807 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data
   [junit4]   2> 2551549 T4807 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data/index/
   [junit4]   2> 2551549 T4807 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard1_replica5] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard1_replica5/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2551552 T4807 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data/index
   [junit4]   2> 2551555 T4767 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2551572 T4807 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191cc979 lockFactory=org.apache.lucene.store.NativeFSLockFactory@193cb3bd),segFN=segments_1,generation=1}
   [junit4]   2> 2551573 T4807 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2551588 T4752 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2551769 T4795 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2551789 T4811 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2551840 T4749 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2551853 T4767 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2551887 T4794 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2551903 T4766 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2551918 T4807 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2551920 T4807 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2551922 T4807 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2551924 T4807 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2551993 T4752 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2551996 T4749 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2552004 T4811 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2552027 T4752 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2552047 T4767 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552055 T4807 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2552056 T4807 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2552057 T4807 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2552058 T4807 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2552059 T4807 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2552098 T4749 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2552103 T4767 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2552116 T4766 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2552118 T4795 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2552125 T4795 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2552141 T4811 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2552149 T4766 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2552174 T4782 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2552249 T4807 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2552255 T4767 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2552257 T4767 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2552257 T4767 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2552279 T4767 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2552280 T4767 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2552281 T4767 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2552282 T4767 oasc.SolrCore.<init> [awholynewcollection_0_shard2_replica4] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard2_replica4/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data/
   [junit4]   2> 2552283 T4767 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2552284 T4767 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data
   [junit4]   2> 2552285 T4767 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data/index/
   [junit4]   2> 2552285 T4767 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard2_replica4] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard2_replica4/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2552316 T4767 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data/index
   [junit4]   2> 2552548 T4766 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552561 T4811 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552570 T4766 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552595 T4795 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552598 T4807 oass.SolrIndexSearcher.<init> Opening Searcher@20840278 main
   [junit4]   2> 2552603 T4807 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2552605 T4807 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2552628 T4794 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2552629 T4782 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2552644 T4794 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2552656 T4749 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552658 T4752 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552671 T4782 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2552676 T4862 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard1_replica5] Registered new searcher Searcher@20840278 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2552687 T4752 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552689 T4749 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552700 T4767 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33b1b80c lockFactory=org.apache.lucene.store.NativeFSLockFactory@70f08a18),segFN=segments_1,generation=1}
   [junit4]   2> 2552712 T4767 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2552715 T4796 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2552741 T4811 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552790 T4767 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2552791 T4767 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2552793 T4767 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2552795 T4767 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2552791 T4807 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard1_replica5
   [junit4]   2> 2552797 T4807 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard1_replica5 address:http://127.0.0.1:50812/i/v collection:awholynewcollection_0 shard:shard1
   [junit4]   2> 2552798 T4807 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2552832 T4795 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552805 T4747 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2552858 T4766 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2552866 T4783 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2552880 T4767 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2552885 T4767 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2552881 T4766 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2552888 T4766 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2552889 T4766 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2552888 T4767 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2552896 T4766 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2552897 T4766 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2552898 T4766 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2552899 T4766 oasc.SolrCore.<init> [awholynewcollection_0_shard3_replica4] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard3_replica4/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data/
   [junit4]   2> 2552900 T4766 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2552903 T4766 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data
   [junit4]   2> 2552896 T4767 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2552906 T4767 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2552915 T4766 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data/index/
   [junit4]   2> 2552916 T4766 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard3_replica4] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard3_replica4/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2552918 T4766 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data/index
   [junit4]   2> 2552923 T4749 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2552926 T4811 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2552938 T4752 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2552968 T4807 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leader_elect/shard1/election
   [junit4]   2> 2553022 T4809 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2553004 T4766 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@403564ca lockFactory=org.apache.lucene.store.NativeFSLockFactory@49e455dc),segFN=segments_1,generation=1}
   [junit4]   2> 2553039 T4766 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2552972 T4752 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553045 T4752 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553045 T4752 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2552972 T4749 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553047 T4749 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553048 T4749 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2553051 T4782 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2552971 T4811 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553056 T4811 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553057 T4811 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2553060 T4752 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553063 T4752 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553062 T4749 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553065 T4749 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553065 T4749 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2553066 T4749 oasc.SolrCore.<init> [awholynewcollection_0_shard4_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard4_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data/
   [junit4]   2> 2553066 T4749 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2553068 T4749 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data
   [junit4]   2> 2553069 T4749 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data/index/
   [junit4]   2> 2553069 T4749 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard4_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard4_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2553064 T4752 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2553072 T4752 oasc.SolrCore.<init> [awholynewcollection_0_shard1_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard1_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data/
   [junit4]   2> 2553072 T4752 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2553074 T4752 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data
   [junit4]   2> 2553074 T4752 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data/index/
   [junit4]   2> 2553075 T4752 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard1_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard1_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2553077 T4811 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553079 T4811 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553079 T4811 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2553080 T4811 oasc.SolrCore.<init> [awholynewcollection_0_shard3_replica5] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard3_replica5/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data/
   [junit4]   2> 2553080 T4811 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2553082 T4811 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data
   [junit4]   2> 2553084 T4749 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data/index
   [junit4]   2> 2553085 T4808 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2553088 T4752 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data/index
   [junit4]   2> 2553106 T4811 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data/index/
   [junit4]   2> 2553107 T4811 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard3_replica5] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard3_replica5/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2553109 T4811 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data/index
   [junit4]   2> 2553111 T4796 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553117 T4796 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2553124 T4783 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553127 T4795 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2553140 T4783 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2553149 T4747 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553157 T4747 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2553161 T4782 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553167 T4809 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553172 T4809 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2553196 T4795 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553198 T4795 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553200 T4795 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2553198 T4752 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@cacdc0f lockFactory=org.apache.lucene.store.NativeFSLockFactory@7671f9f7),segFN=segments_1,generation=1}
   [junit4]   2> 2553206 T4752 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2553216 T4749 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@9161090 lockFactory=org.apache.lucene.store.NativeFSLockFactory@27d919f),segFN=segments_1,generation=1}
   [junit4]   2> 2553217 T4749 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2553237 T4811 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@47fcc3de lockFactory=org.apache.lucene.store.NativeFSLockFactory@52f20eaa),segFN=segments_1,generation=1}
   [junit4]   2> 2553238 T4811 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2553241 T4766 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2553242 T4766 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2553273 T4766 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2553287 T4795 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553288 T4795 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553290 T4795 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2553291 T4795 oasc.SolrCore.<init> [awholynewcollection_0_shard1_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard1_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data/
   [junit4]   2> 2553292 T4795 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2553293 T4795 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data
   [junit4]   2> 2553294 T4795 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data/index/
   [junit4]   2> 2553294 T4795 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard1_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard1_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2553289 T4794 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553297 T4795 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data/index
   [junit4]   2> 2553290 T4766 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2553322 T4766 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2553322 T4766 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2553323 T4766 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2553324 T4766 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2553325 T4766 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2553326 T4794 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553298 T4807 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 2553338 T4794 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2553343 T4794 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553346 T4794 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553347 T4794 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2553348 T4780 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2553350 T4807 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Waiting until we see more replicas up for shard shard1: total=5 found=1 timeoutin=179993
   [junit4]   2> 2553355 T4749 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2553355 T4749 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2553363 T4748 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2553399 T4749 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2553401 T4749 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2553403 T4749 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2553404 T4749 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2553405 T4749 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2553407 T4749 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2553421 T4752 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2553423 T4752 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2553426 T4752 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2553615 T4752 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2553617 T4752 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2553618 T4752 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2553613 T4809 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553620 T4752 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2553622 T4752 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2553623 T4752 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2553601 T4855 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2553594 T4766 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2553590 T4748 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553582 T4796 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553569 T4768 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2553568 T4780 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553558 T4795 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@35729ebf lockFactory=org.apache.lucene.store.NativeFSLockFactory@13bc8d02),segFN=segments_1,generation=1}
   [junit4]   2> 2553642 T4795 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2553554 T4784 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2553527 T4749 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2553648 T4747 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553517 T4794 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553649 T4794 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553650 T4794 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2553486 T4811 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2553652 T4811 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2553655 T4811 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2553657 T4811 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2553658 T4811 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2553659 T4811 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2553660 T4811 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2553473 T4767 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2553454 T4783 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553448 T4808 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553668 T4749 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2553440 T4782 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2553672 T4808 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2553661 T4811 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2553677 T4811 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2553651 T4794 oasc.SolrCore.<init> [awholynewcollection_0_shard2_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard2_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data/
   [junit4]   2> 2553681 T4794 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2553681 T4811 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2553683 T4794 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data
   [junit4]   2> 2553683 T4794 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data/index/
   [junit4]   2> 2553684 T4794 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard2_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard2_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2553691 T4784 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553797 T4811 oass.SolrIndexSearcher.<init> Opening Searcher@26d9a2b main
   [junit4]   2> 2553802 T4811 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2553803 T4811 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2553762 T4749 oass.SolrIndexSearcher.<init> Opening Searcher@49f9e0ea main
   [junit4]   2> 2553811 T4749 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2553813 T4749 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2553816 T4795 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2553817 T4795 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2553745 T4747 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553744 T4748 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2553719 T4780 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2553715 T4796 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553712 T4782 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553838 T4782 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553839 T4782 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2553711 T4809 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553711 T4783 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553696 T4767 oass.SolrIndexSearcher.<init> Opening Searcher@1d6c0286 main
   [junit4]   2> 2553696 T4794 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data/index
   [junit4]   2> 2553849 T4767 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2553851 T4767 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2553693 T4752 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2553847 T4782 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2553879 T4782 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2553881 T4782 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2553882 T4864 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard4_replica2] Registered new searcher Searcher@49f9e0ea main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2553819 T4795 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2553899 T4795 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2553900 T4795 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2553901 T4795 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2553902 T4795 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2553905 T4795 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2553906 T4795 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2553908 T4808 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2553897 T4749 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard4_replica2
   [junit4]   2> 2553910 T4749 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard4_replica2 address:http://127.0.0.1:50799/i/v collection:awholynewcollection_0 shard:shard4
   [junit4]   2> 2553912 T4749 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2553882 T4782 oasc.SolrCore.<init> [awholynewcollection_0_shard5_replica3] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard5_replica3/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data/
   [junit4]   2> 2553922 T4782 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2553924 T4782 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data
   [junit4]   2> 2553878 T4766 oass.SolrIndexSearcher.<init> Opening Searcher@7dd2b92d main
   [junit4]   2> 2553928 T4766 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2553929 T4766 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2553931 T4855 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2553926 T4794 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@738673ee lockFactory=org.apache.lucene.store.NativeFSLockFactory@6ad5266b),segFN=segments_1,generation=1}
   [junit4]   2> 2553935 T4782 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data/index/
   [junit4]   2> 2553938 T4794 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2553906 T4865 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard3_replica5] Registered new searcher Searcher@26d9a2b main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2553903 T4863 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard2_replica4] Registered new searcher Searcher@1d6c0286 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2553939 T4782 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard5_replica3] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard5_replica3/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2553945 T4782 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data/index
   [junit4]   2> 2554014 T4767 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard2_replica4
   [junit4]   2> 2554015 T4767 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard2_replica4 address:http://127.0.0.1:50803/i/v collection:awholynewcollection_0 shard:shard2
   [junit4]   2> 2554016 T4767 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2554017 T4794 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2554018 T4794 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2554020 T4794 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2554023 T4794 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2554026 T4866 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard3_replica4] Registered new searcher Searcher@7dd2b92d main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2554034 T4794 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2554036 T4794 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2554037 T4794 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2554039 T4766 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard3_replica4
   [junit4]   2> 2554040 T4766 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard3_replica4 address:http://127.0.0.1:50803/i/v collection:awholynewcollection_0 shard:shard3
   [junit4]   2> 2554041 T4766 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2554063 T4811 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard3_replica5
   [junit4]   2> 2554064 T4811 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard3_replica5 address:http://127.0.0.1:50812/i/v collection:awholynewcollection_0 shard:shard3
   [junit4]   2> 2554065 T4811 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2554076 T4809 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554082 T4794 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2554106 T4794 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2554096 T4783 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554094 T4809 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554110 T4809 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554111 T4809 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554117 T4783 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554118 T4783 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554119 T4783 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554125 T4809 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554128 T4783 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554128 T4783 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554129 T4783 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2554130 T4783 oasc.SolrCore.<init> [awholynewcollection_0_shard2_replica3] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard2_replica3/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data/
   [junit4]   2> 2554131 T4783 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2554132 T4783 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data
   [junit4]   2> 2554133 T4783 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data/index/
   [junit4]   2> 2554133 T4783 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard2_replica3] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard2_replica3/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2554139 T4747 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554112 T4855 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2554147 T4784 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2554150 T4783 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data/index
   [junit4]   2> 2554128 T4809 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554127 T4795 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2554153 T4809 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2554155 T4809 oasc.SolrCore.<init> [awholynewcollection_0_shard4_replica5] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard4_replica5/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data/
   [junit4]   2> 2554155 T4809 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2554159 T4768 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2554166 T4768 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2554177 T4809 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data
   [junit4]   2> 2554179 T4809 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data/index/
   [junit4]   2> 2554180 T4809 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard4_replica5] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard4_replica5/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2554175 T4747 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554185 T4747 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554186 T4747 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554175 T4748 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2554193 T4809 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data/index
   [junit4]   2> 2554202 T4808 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2554211 T4752 oass.SolrIndexSearcher.<init> Opening Searcher@7877e960 main
   [junit4]   2> 2554216 T4752 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2554220 T4752 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2554228 T4766 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leader_elect/shard3/election
   [junit4]   2> 2554232 T4747 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554234 T4747 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554235 T4747 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2554237 T4747 oasc.SolrCore.<init> [awholynewcollection_0_shard2_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard2_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data/
   [junit4]   2> 2554238 T4747 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2554237 T4808 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554241 T4747 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data
   [junit4]   2> 2554252 T4747 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data/index/
   [junit4]   2> 2554253 T4747 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard2_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard2_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2554253 T4811 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leader_elect/shard3/election
   [junit4]   2> 2554268 T4809 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7ce43eb8 lockFactory=org.apache.lucene.store.NativeFSLockFactory@71c2fd3a),segFN=segments_1,generation=1}
   [junit4]   2> 2554269 T4809 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2554272 T4747 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data/index
   [junit4]   2> 2554274 T4867 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard1_replica2] Registered new searcher Searcher@7877e960 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2554282 T4752 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard1_replica2
   [junit4]   2> 2554288 T4752 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard1_replica2 address:http://127.0.0.1:50799/i/v collection:awholynewcollection_0 shard:shard1
   [junit4]   2> 2554288 T4752 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2554291 T4783 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@72edb13a lockFactory=org.apache.lucene.store.NativeFSLockFactory@65db621b),segFN=segments_1,generation=1}
   [junit4]   2> 2554315 T4783 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2554299 T4796 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554327 T4796 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554331 T4796 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554332 T4796 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554340 T4796 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554342 T4796 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554343 T4796 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2554344 T4796 oasc.SolrCore.<init> [awholynewcollection_0_shard5_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard5_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data/
   [junit4]   2> 2554345 T4796 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2554347 T4796 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data
   [junit4]   2> 2554349 T4782 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2d9bb8 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1bf31534),segFN=segments_1,generation=1}
   [junit4]   2> 2554350 T4782 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2554360 T4796 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data/index/
   [junit4]   2> 2554361 T4796 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard5_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard5_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2554363 T4796 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data/index
   [junit4]   2> 2554367 T4795 oass.SolrIndexSearcher.<init> Opening Searcher@52a8cad main
   [junit4]   2> 2554370 T4795 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2554372 T4795 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2554381 T4868 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard1_replica1] Registered new searcher Searcher@52a8cad main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2554382 T4808 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554383 T4808 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554384 T4808 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2554389 T4749 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leader_elect/shard4/election
   [junit4]   2> 2554401 T4782 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2554403 T4782 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2554424 T4809 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2554425 T4809 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2554430 T4855 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2554433 T4780 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2554435 T4767 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leader_elect/shard2/election
   [junit4]   2> 2554456 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2554460 T4795 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard1_replica1
   [junit4]   2> 2554461 T4795 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard1_replica1 address:http://127.0.0.1:50809/i/v collection:awholynewcollection_0 shard:shard1
   [junit4]   2> 2554462 T4795 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2554466 T4747 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@3e473239 lockFactory=org.apache.lucene.store.NativeFSLockFactory@16e642b4),segFN=segments_1,generation=1}
   [junit4]   2> 2554468 T4747 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2554470 T4783 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2554476 T4766 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard3
   [junit4]   2> 2554484 T4784 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2554493 T4831 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2554494 T4761 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2554496 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2554500 T4747 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2554559 T4747 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2554562 T4747 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2554563 T4747 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2554564 T4747 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2554565 T4747 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2554565 T4796 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@6daa6594 lockFactory=org.apache.lucene.store.NativeFSLockFactory@6158b457),segFN=segments_1,generation=1}
   [junit4]   2> 2554566 T4796 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2554471 T4808 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2554577 T4808 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2554578 T4808 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2554580 T4808 oasc.SolrCore.<init> [awholynewcollection_0_shard2_replica5] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard2_replica5/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data/
   [junit4]   2> 2554580 T4808 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2554582 T4808 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data
   [junit4]   2> 2554583 T4808 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data/index/
   [junit4]   2> 2554583 T4808 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard2_replica5] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard2_replica5/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2554565 T4747 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2554586 T4747 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2554587 T4747 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2554540 T4748 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2554537 T4780 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2554509 T4783 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2554509 T4794 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2554507 T4809 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2554507 T4782 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2554505 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2554505 T4766 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Waiting until we see more replicas up for shard shard3: total=5 found=1 timeoutin=179995
   [junit4]   2> 2554505 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2554502 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2554494 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2554494 T4833 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2554640 T4782 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2554646 T4782 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2554646 T4782 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2554639 T4809 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2554604 T4783 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2554652 T4783 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2554659 T4796 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2554660 T4796 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2554663 T4796 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2554601 T4808 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data/index
   [junit4]   2> 2554660 T4783 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2554671 T4783 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2554672 T4783 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2554674 T4783 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2554686 T4783 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2554649 T4809 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2554694 T4809 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2554695 T4809 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2554647 T4782 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2554699 T4782 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2554702 T4782 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2554697 T4809 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2554675 T4794 oass.SolrIndexSearcher.<init> Opening Searcher@74797957 main
   [junit4]   2> 2554708 T4809 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2554714 T4784 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2554756 T4796 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2554759 T4808 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7fdc3e74 lockFactory=org.apache.lucene.store.NativeFSLockFactory@2c0ea18),segFN=segments_1,generation=1}
   [junit4]   2> 2554761 T4808 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2554764 T4749 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2554772 T4794 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2554774 T4794 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2554790 T4869 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard2_replica1] Registered new searcher Searcher@74797957 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2554761 T4796 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2554795 T4796 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2554796 T4796 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2554827 T4796 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2554830 T4796 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2554793 T4794 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard2_replica1
   [junit4]   2> 2554833 T4794 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard2_replica1 address:http://127.0.0.1:50809/i/v collection:awholynewcollection_0 shard:shard2
   [junit4]   2> 2554834 T4794 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2554844 T4767 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 2554857 T4749 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Waiting until we see more replicas up for shard shard4: total=5 found=1 timeoutin=179986
   [junit4]   2> 2554869 T4808 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2554870 T4808 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2554871 T4808 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2554877 T4808 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2554878 T4808 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2554879 T4808 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2554879 T4808 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2554881 T4808 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2554882 T4808 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2554892 T4798 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2554894 T4784 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555089 T4796 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2554893 T4780 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555107 T4767 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Waiting until we see more replicas up for shard shard2: total=5 found=1 timeoutin=179792
   [junit4]   2> 2555109 T4780 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555110 T4780 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555113 T4780 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555114 T4784 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555115 T4784 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555116 T4784 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555117 T4782 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555125 T4809 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555136 T4784 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555137 T4784 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555138 T4784 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2555143 T4784 oasc.SolrCore.<init> [awholynewcollection_0_shard3_replica3] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard3_replica3/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data/
   [junit4]   2> 2555144 T4784 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2555146 T4784 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data
   [junit4]   2> 2555147 T4784 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data/index/
   [junit4]   2> 2555147 T4784 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard3_replica3] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard3_replica3/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2555138 T4747 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555150 T4780 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555152 T4780 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555153 T4780 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2555157 T4780 oasc.SolrCore.<init> [awholynewcollection_0_shard1_replica3] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard1_replica3/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data/
   [junit4]   2> 2555158 T4780 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2555159 T4784 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data/index
   [junit4]   2> 2555155 T4768 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2555168 T4780 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data
   [junit4]   2> 2555180 T4780 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data/index/
   [junit4]   2> 2555181 T4780 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard1_replica3] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard1_replica3/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2555166 T4747 oass.SolrIndexSearcher.<init> Opening Searcher@38c9d98d main
   [junit4]   2> 2555184 T4782 oass.SolrIndexSearcher.<init> Opening Searcher@4cc04823 main
   [junit4]   2> 2555187 T4782 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555189 T4782 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2555166 T4809 oass.SolrIndexSearcher.<init> Opening Searcher@73f76d11 main
   [junit4]   2> 2555202 T4809 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555203 T4809 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2555184 T4747 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555208 T4747 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2555215 T4872 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard4_replica5] Registered new searcher Searcher@73f76d11 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2555181 T4855 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2555220 T4780 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data/index
   [junit4]   2> 2555222 T4809 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard4_replica5
   [junit4]   2> 2555223 T4809 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard4_replica5 address:http://127.0.0.1:50812/i/v collection:awholynewcollection_0 shard:shard4
   [junit4]   2> 2555224 T4809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2555175 T4783 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555234 T4873 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard2_replica2] Registered new searcher Searcher@38c9d98d main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2555175 T4784 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@3e815ecb lockFactory=org.apache.lucene.store.NativeFSLockFactory@7a3c923a),segFN=segments_1,generation=1}
   [junit4]   2> 2555240 T4748 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555207 T4808 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555245 T4747 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard2_replica2
   [junit4]   2> 2555249 T4747 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard2_replica2 address:http://127.0.0.1:50799/i/v collection:awholynewcollection_0 shard:shard2
   [junit4]   2> 2555250 T4747 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2555251 T4796 oass.SolrIndexSearcher.<init> Opening Searcher@17260668 main
   [junit4]   2> 2555240 T4784 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2555259 T4871 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard5_replica3] Registered new searcher Searcher@4cc04823 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2555263 T4782 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard5_replica3
   [junit4]   2> 2555265 T4796 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555274 T4796 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2555264 T4782 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard5_replica3 address:http://127.0.0.1:50806/i/v collection:awholynewcollection_0 shard:shard5
   [junit4]   2> 2555277 T4782 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2555280 T4748 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555282 T4748 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555283 T4748 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555309 T4748 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555311 T4748 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555312 T4748 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2555313 T4748 oasc.SolrCore.<init> [awholynewcollection_0_shard3_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard3_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data/
   [junit4]   2> 2555315 T4748 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2555322 T4765 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2555339 T4783 oass.SolrIndexSearcher.<init> Opening Searcher@7c772317 main
   [junit4]   2> 2555388 T4870 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard5_replica1] Registered new searcher Searcher@17260668 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2555340 T4748 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data
   [junit4]   2> 2555390 T4748 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data/index/
   [junit4]   2> 2555391 T4748 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard3_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard3_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2555384 T4784 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2555394 T4784 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2555395 T4784 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2555396 T4768 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2555411 T4748 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data/index
   [junit4]   2> 2555382 T4808 oass.SolrIndexSearcher.<init> Opening Searcher@7a333abb main
   [junit4]   2> 2555418 T4808 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555420 T4808 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2555381 T4780 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@109a0a9f lockFactory=org.apache.lucene.store.NativeFSLockFactory@37ee5bc2),segFN=segments_1,generation=1}
   [junit4]   2> 2555423 T4780 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2555379 T4796 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard5_replica1
   [junit4]   2> 2555426 T4796 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard5_replica1 address:http://127.0.0.1:50809/i/v collection:awholynewcollection_0 shard:shard5
   [junit4]   2> 2555427 T4796 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2555397 T4784 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2555437 T4784 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2555438 T4784 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2555439 T4784 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2555443 T4875 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard2_replica5] Registered new searcher Searcher@7a333abb main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2555447 T4784 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2555456 T4783 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555458 T4783 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2555472 T4782 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leader_elect/shard5/election
   [junit4]   2> 2555476 T4874 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard2_replica3] Registered new searcher Searcher@7c772317 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2555484 T4748 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7ded954d lockFactory=org.apache.lucene.store.NativeFSLockFactory@72ce036),segFN=segments_1,generation=1}
   [junit4]   2> 2555489 T4748 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2555488 T4768 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555495 T4783 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard2_replica3
   [junit4]   2> 2555498 T4783 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard2_replica3 address:http://127.0.0.1:50806/i/v collection:awholynewcollection_0 shard:shard2
   [junit4]   2> 2555499 T4783 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2555504 T4808 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard2_replica5
   [junit4]   2> 2555507 T4808 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard2_replica5 address:http://127.0.0.1:50812/i/v collection:awholynewcollection_0 shard:shard2
   [junit4]   2> 2555508 T4780 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2555509 T4780 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2555511 T4780 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2555513 T4780 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2555524 T4780 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2555525 T4780 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2555526 T4780 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2555527 T4780 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2555528 T4780 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2555533 T4768 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555534 T4768 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555535 T4768 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555549 T4798 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2555553 T4768 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555553 T4768 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555554 T4768 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2555555 T4768 oasc.SolrCore.<init> [awholynewcollection_0_shard4_replica4] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard4_replica4/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data/
   [junit4]   2> 2555556 T4768 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2555557 T4768 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data
   [junit4]   2> 2555558 T4768 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data/index/
   [junit4]   2> 2555558 T4768 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard4_replica4] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard4_replica4/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2555560 T4750 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2555496 T4784 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2555570 T4768 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data/index
   [junit4]   2> 2555569 T4765 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2555563 T4798 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2555508 T4808 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2555593 T4784 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555604 T4855 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555612 T4855 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555613 T4855 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555614 T4855 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2555621 T4782 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard5
   [junit4]   2> 2555622 T4855 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2555623 T4855 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2555632 T4855 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2555633 T4855 oasc.SolrCore.<init> [awholynewcollection_0_shard5_replica5] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard5_replica5/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data/
   [junit4]   2> 2555634 T4855 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2555637 T4855 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data
   [junit4]   2> 2555638 T4855 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data/index/
   [junit4]   2> 2555665 T4782 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Waiting until we see more replicas up for shard shard5: total=5 found=1 timeoutin=179995
   [junit4]   2> 2555637 T4748 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2555669 T4748 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2555671 T4748 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2555674 T4748 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2555677 T4748 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2555678 T4748 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2555680 T4748 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2555706 T4748 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2555711 T4748 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2555668 T4768 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@313a0ac7 lockFactory=org.apache.lucene.store.NativeFSLockFactory@71f736af),segFN=segments_1,generation=1}
   [junit4]   2> 2555725 T4768 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2555665 T4855 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard5_replica5] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749/awholynewcollection_0_shard5_replica5/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2555724 T4784 oass.SolrIndexSearcher.<init> Opening Searcher@70e439a4 main
   [junit4]   2> 2555736 T4784 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555745 T4784 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2555704 T4770 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2555689 T4780 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555772 T4876 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard3_replica3] Registered new searcher Searcher@70e439a4 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2555688 T4750 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2555682 T4765 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2555746 T4855 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data/index
   [junit4]   2> 2555745 T4768 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2555789 T4768 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2555791 T4768 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2555792 T4768 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2555792 T4784 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard3_replica3
   [junit4]   2> 2555793 T4784 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard3_replica3 address:http://127.0.0.1:50806/i/v collection:awholynewcollection_0 shard:shard3
   [junit4]   2> 2555795 T4784 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2555786 T4780 oass.SolrIndexSearcher.<init> Opening Searcher@3fd66f46 main
   [junit4]   2> 2555807 T4780 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555808 T4780 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2555821 T4748 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555793 T4768 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2555824 T4768 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2555826 T4768 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2555829 T4768 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2555831 T4768 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2555822 T4750 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2555856 T4768 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2555888 T4798 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2555866 T4770 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2555867 T4855 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@79dae424 lockFactory=org.apache.lucene.store.NativeFSLockFactory@6df2a863),segFN=segments_1,generation=1}
   [junit4]   2> 2555898 T4855 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2555905 T4877 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard1_replica3] Registered new searcher Searcher@3fd66f46 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2555908 T4780 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard1_replica3
   [junit4]   2> 2555909 T4780 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard1_replica3 address:http://127.0.0.1:50806/i/v collection:awholynewcollection_0 shard:shard1
   [junit4]   2> 2555909 T4780 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2555923 T4855 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2555924 T4855 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2555932 T4855 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2555934 T4855 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2555936 T4855 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2555938 T4855 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2555940 T4855 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2555974 T4770 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2555979 T4748 oass.SolrIndexSearcher.<init> Opening Searcher@126d186b main
   [junit4]   2> 2555981 T4748 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2555982 T4748 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2556001 T4878 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard3_replica2] Registered new searcher Searcher@126d186b main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2556013 T4797 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2556037 T4768 oass.SolrIndexSearcher.<init> Opening Searcher@7b8f9eab main
   [junit4]   2> 2556041 T4768 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2556042 T4768 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2556047 T4798 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556065 T4748 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard3_replica2
   [junit4]   2> 2556066 T4748 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard3_replica2 address:http://127.0.0.1:50799/i/v collection:awholynewcollection_0 shard:shard3
   [junit4]   2> 2556067 T4748 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2556079 T4879 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard4_replica4] Registered new searcher Searcher@7b8f9eab main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2556090 T4855 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2556091 T4855 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2556138 T4798 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2556185 T4769 oasha.CollectionsHandler.handleCreateAction Creating Collection : numShards=5&maxShardsPerNode=6&name=awholynewcollection_0&replicationFactor=5&action=CREATE&wt=javabin&version=2
   [junit4]   2> 2556202 T4798 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2556209 T4798 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2556211 T4798 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2556215 T4768 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard4_replica4
   [junit4]   2> 2556216 T4768 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard4_replica4 address:http://127.0.0.1:50803/i/v collection:awholynewcollection_0 shard:shard4
   [junit4]   2> 2556218 T4768 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2556179 T4779 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 2556209 T4767 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 2556228 T4767 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2556228 T4767 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/
   [junit4]   2> 2556208 T4756 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work state: SyncConnected type NodeChildrenChanged
   [junit4]   2> 2556205 T4765 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556229 T4767 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard2_replica4 url=http://127.0.0.1:50803/i/v START replicas=[http://127.0.0.1:50809/i/v/awholynewcollection_0_shard2_replica1/, http://127.0.0.1:50799/i/v/awholynewcollection_0_shard2_replica2/, http://127.0.0.1:50806/i/v/awholynewcollection_0_shard2_replica3/, http://127.0.0.1:50812/i/v/awholynewcollection_0_shard2_replica5/] nUpdates=100
   [junit4]   2> 2556232 T4798 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2556240 T4750 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556237 T4767 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard2_replica4 url=http://127.0.0.1:50803/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2556245 T4767 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2556246 T4767 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2556246 T4767 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/ shard2
   [junit4]   2> 2556247 T4767 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leaders/shard2
   [junit4]   2> 2556254 T4798 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2556257 T4798 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2556263 T4798 oasc.SolrCore.<init> [awholynewcollection_0_shard4_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard4_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data/
   [junit4]   2> 2556271 T4798 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2556282 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2556283 T4798 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data
   [junit4]   2> 2556292 T4798 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data/index/
   [junit4]   2> 2556298 T4798 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard4_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard4_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2556334 T4765 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556346 T4750 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> ASYNC  NEW_CORE C817 name=awholynewcollection_0_shard2_replica3 org.apache.solr.core.SolrCore@7c62b1db url=http://127.0.0.1:50806/i/v/awholynewcollection_0_shard2_replica3 node=127.0.0.1:50806_i%2Fv C817_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica3 props:{state=down, core=awholynewcollection_0_shard2_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2556361 T4781 C817 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=11 
   [junit4]   2> ASYNC  NEW_CORE C818 name=awholynewcollection_0_shard2_replica1 org.apache.solr.core.SolrCore@7681485 url=http://127.0.0.1:50809/i/v/awholynewcollection_0_shard2_replica1 node=127.0.0.1:50809_i%2Fv C818_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica1 props:{state=down, core=awholynewcollection_0_shard2_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2556362 T4793 C818 P50809 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica1] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=11 
   [junit4]   2> ASYNC  NEW_CORE C819 name=awholynewcollection_0_shard2_replica2 org.apache.solr.core.SolrCore@2dd08a87 url=http://127.0.0.1:50799/i/v/awholynewcollection_0_shard2_replica2 node=127.0.0.1:50799_i%2Fv C819_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica2 props:{state=down, core=awholynewcollection_0_shard2_replica2, node_name=127.0.0.1:50799_i%2Fv, base_url=http://127.0.0.1:50799/i/v}
   [junit4]   2> 2556365 T4751 C819 P50799 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica2] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=19 
   [junit4]   2> 2556368 T4855 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2556409 T4798 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data/index
   [junit4]   2> ASYNC  NEW_CORE C820 name=awholynewcollection_0_shard2_replica5 org.apache.solr.core.SolrCore@a3d63c8 url=http://127.0.0.1:50812/i/v/awholynewcollection_0_shard2_replica5 node=127.0.0.1:50812_i%2Fv C820_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica5 props:{state=down, core=awholynewcollection_0_shard2_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2556409 T4810 C820 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=5 
   [junit4]   2> 2556416 T4797 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2556419 T4779 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 2556429 T4779 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2556460 T4798 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5430818f lockFactory=org.apache.lucene.store.NativeFSLockFactory@568b9093),segFN=segments_1,generation=1}
   [junit4]   2> 2556461 T4798 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2556444 T4765 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2556441 T4855 oass.SolrIndexSearcher.<init> Opening Searcher@1d7741f5 main
   [junit4]   2> 2556439 T4797 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 2556435 T4770 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556490 T4765 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2556491 T4765 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2556491 T4765 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2556432 T4750 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2556491 T4855 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2556498 T4855 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2556512 T4798 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2556525 T4798 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2556514 T4750 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2556527 T4750 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2556528 T4750 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2556524 T4855 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard5_replica5
   [junit4]   2> 2556529 T4855 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard5_replica5 address:http://127.0.0.1:50812/i/v collection:awholynewcollection_0 shard:shard5
   [junit4]   2> 2556521 T4885 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard5_replica5] Registered new searcher Searcher@1d7741f5 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2556518 T4765 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2556532 T4765 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2556533 T4765 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2556534 T4765 oasc.SolrCore.<init> [awholynewcollection_0_shard5_replica4] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard5_replica4/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data/
   [junit4]   2> 2556535 T4765 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2556537 T4765 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data
   [junit4]   2> 2556538 T4765 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data/index/
   [junit4]   2> 2556540 T4765 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard5_replica4] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard5_replica4/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2556531 T4855 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2556538 T4750 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2556545 T4750 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2556546 T4750 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2556548 T4750 oasc.SolrCore.<init> [awholynewcollection_0_shard5_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard5_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data/
   [junit4]   2> 2556548 T4750 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2556551 T4750 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data
   [junit4]   2> 2556553 T4750 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data/index/
   [junit4]   2> 2556555 T4750 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard5_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389/awholynewcollection_0_shard5_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2556554 T4765 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data/index
   [junit4]   2> 2556569 T4750 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data/index
   [junit4]   2> 2556570 T4770 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556603 T4798 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2556605 T4798 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2556607 T4798 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2556608 T4798 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2556609 T4770 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2556620 T4770 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2556621 T4770 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2556622 T4770 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2556632 T4765 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2b87fc72 lockFactory=org.apache.lucene.store.NativeFSLockFactory@65b811ae),segFN=segments_1,generation=1}
   [junit4]   2> 2556634 T4765 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2556640 T4761 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2556641 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2556642 T4833 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2556643 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2556661 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2556666 T4770 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2556667 T4770 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2556668 T4770 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2556670 T4770 oasc.SolrCore.<init> [awholynewcollection_0_shard1_replica4] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard1_replica4/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data/
   [junit4]   2> 2556671 T4770 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2556676 T4767 oasc.ZkController.register We are http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/
   [junit4]   2> 2556678 T4767 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 2556678 T4767 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica4 state=active
   [junit4]   2> 2556688 T4765 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2556689 T4765 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2556691 T4765 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2556693 T4765 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2556609 T4798 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2556642 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2556640 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2556640 T4831 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2556700 T4750 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2b1f5198 lockFactory=org.apache.lucene.store.NativeFSLockFactory@3f35b24b),segFN=segments_1,generation=1}
   [junit4]   2> 2556753 T4797 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556751 T4794 oasc.ZkController.register We are http://127.0.0.1:50809/i/v/awholynewcollection_0_shard2_replica1/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/
   [junit4]   2> 2556761 T4794 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard2_replica1 baseURL=http://127.0.0.1:50809/i/v
   [junit4]   2> 2556762 T4794 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard2_replica1
   [junit4]   2> 2556762 T4794 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2556741 T4798 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2556736 T4808 oasc.ZkController.register We are http://127.0.0.1:50812/i/v/awholynewcollection_0_shard2_replica5/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/
   [junit4]   2> 2556769 T4798 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2556777 T4794 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2556780 T4794 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard2&name=awholynewcollection_0_shard2_replica1&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=30629 
   [junit4]   2> 2556714 T4767 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2556802 T4767 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard2&name=awholynewcollection_0_shard2_replica4&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=30630 
   [junit4]   2> 2556812 T4798 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2556736 T4765 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2556817 T4765 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2556819 T4765 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2556822 T4765 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2556823 T4765 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2556731 T4770 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data
   [junit4]   2> 2556729 T4783 oasc.ZkController.register We are http://127.0.0.1:50806/i/v/awholynewcollection_0_shard2_replica3/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/
   [junit4]   2> 2556831 T4783 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard2_replica3 baseURL=http://127.0.0.1:50806/i/v
   [junit4]   2> 2556832 T4783 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard2_replica3
   [junit4]   2> 2556833 T4783 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2556769 T4808 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard2_replica5 baseURL=http://127.0.0.1:50812/i/v
   [junit4]   2> 2556841 T4808 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard2_replica5
   [junit4]   2> 2556845 T4808 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2556765 T4886 C818 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard2_replica1 recoveringAfterStartup=true
   [junit4]   2> 2556848 T4886 C818 P50809 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2556849 T4886 C818 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica1 state=recovering
   [junit4]   2> 2556853 T4888 C820 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard2_replica5 recoveringAfterStartup=true
   [junit4]   2> 2556854 T4888 C820 P50812 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2556854 T4888 C820 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica5 state=recovering
   [junit4]   2> 2556754 T4750 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2556856 T4783 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2556847 T4779 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556845 T4887 C817 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard2_replica3 recoveringAfterStartup=true
   [junit4]   2> 2556864 T4887 C817 P50806 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2556865 T4887 C817 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica3 state=recovering
   [junit4]   2> 2556834 T4765 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2556878 T4783 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard2&name=awholynewcollection_0_shard2_replica3&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=30706 
   [junit4]   2> 2556832 T4770 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data/index/
   [junit4]   2> 2556867 T4888 C820 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2556859 T4808 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2556859 T4886 C818 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2556883 T4750 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2556884 T4750 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2556885 T4750 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2556886 T4750 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2556893 T4750 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2556896 T4750 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2556904 T4750 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2556912 T4750 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2556917 T4750 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2556881 T4887 C817 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2556930 T4747 oasc.ZkController.register We are http://127.0.0.1:50799/i/v/awholynewcollection_0_shard2_replica2/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/
   [junit4]   2> 2556932 T4747 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard2_replica2 baseURL=http://127.0.0.1:50799/i/v
   [junit4]   2> 2556933 T4747 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard2_replica2
   [junit4]   2> 2556933 T4747 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2556936 T4889 C819 P50799 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard2_replica2 recoveringAfterStartup=true
   [junit4]   2> 2556937 T4889 C819 P50799 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2556937 T4889 C819 P50799 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica2 state=recovering
   [junit4]   2> 2556881 T4770 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard1_replica4] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008/awholynewcollection_0_shard1_replica4/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2556900 T4808 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard2&name=awholynewcollection_0_shard2_replica5&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=30720 
   [junit4]   2> 2556974 T4889 C819 P50799 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2556977 T4770 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data/index
   [junit4]   2> 2557008 T4890 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node9, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2556887 T4797 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2557038 T4893 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node8, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2557031 T4765 oass.SolrIndexSearcher.<init> Opening Searcher@1248fce2 main
   [junit4]   2> 2557042 T4765 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2557046 T4765 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2557030 T4750 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2557052 T4798 oass.SolrIndexSearcher.<init> Opening Searcher@4cbe6a69 main
   [junit4]   2> 2557059 T4798 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2557062 T4798 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2557023 T4779 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 2556997 T4858 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node7, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2556987 T4767 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node6, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2556963 T4747 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2557082 T4747 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard2&name=awholynewcollection_0_shard2_replica2&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=30923 
   [junit4]   2> 2557101 T4770 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@72bf5d lockFactory=org.apache.lucene.store.NativeFSLockFactory@38e1d2a9),segFN=segments_1,generation=1}
   [junit4]   2> 2557108 T4892 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard5_replica4] Registered new searcher Searcher@1248fce2 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2557118 T4891 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard4_replica1] Registered new searcher Searcher@4cbe6a69 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2557112 T4770 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2557133 T4797 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2557142 T4779 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2557140 T4798 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard4_replica1
   [junit4]   2> 2557146 T4798 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard4_replica1 address:http://127.0.0.1:50809/i/v collection:awholynewcollection_0 shard:shard4
   [junit4]   2> 2557147 T4798 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2557137 T4765 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard5_replica4
   [junit4]   2> 2557149 T4765 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard5_replica4 address:http://127.0.0.1:50803/i/v collection:awholynewcollection_0 shard:shard5
   [junit4]   2> 2557149 T4765 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2557151 T4770 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2557153 T4770 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2557155 T4770 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2557158 T4770 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2557169 T4770 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2557170 T4770 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2557171 T4770 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2557174 T4770 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2557161 T4750 oass.SolrIndexSearcher.<init> Opening Searcher@4a769e74 main
   [junit4]   2> 2557200 T4750 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2557205 T4750 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2557186 T4770 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2557182 T4797 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2557215 T4797 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2557218 T4797 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2557182 T4779 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2557222 T4779 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2557224 T4779 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 2557229 T4894 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard5_replica2] Registered new searcher Searcher@4a769e74 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2557236 T4779 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2557238 T4779 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2557239 T4779 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2557240 T4779 oasc.SolrCore.<init> [awholynewcollection_0_shard4_replica3] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard4_replica3/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data/
   [junit4]   2> 2557240 T4779 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2557245 T4779 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data
   [junit4]   2> 2557246 T4779 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data/index/
   [junit4]   2> 2557247 T4779 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard4_replica3] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194/awholynewcollection_0_shard4_replica3/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2557249 T4750 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard5_replica2
   [junit4]   2> 2557250 T4797 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 2557253 T4797 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 2557250 T4750 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard5_replica2 address:http://127.0.0.1:50799/i/v collection:awholynewcollection_0 shard:shard5
   [junit4]   2> 2557255 T4750 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2557257 T4797 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 2557259 T4797 oasc.SolrCore.<init> [awholynewcollection_0_shard3_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard3_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data/
   [junit4]   2> 2557261 T4797 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@46644e4b
   [junit4]   2> 2557264 T4797 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data
   [junit4]   2> 2557266 T4797 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data/index/
   [junit4]   2> 2557267 T4797 oasc.SolrCore.initIndex WARN [awholynewcollection_0_shard3_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558/awholynewcollection_0_shard3_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 2557260 T4779 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data/index
   [junit4]   2> 2557280 T4797 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data/index
   [junit4]   2> 2557289 T4770 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2557314 T4797 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@77864c78 lockFactory=org.apache.lucene.store.NativeFSLockFactory@104fae7b),segFN=segments_1,generation=1}
   [junit4]   2> 2557315 T4797 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2557322 T4770 oass.SolrIndexSearcher.<init> Opening Searcher@190f9e4a main
   [junit4]   2> 2557330 T4770 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2557331 T4770 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2557335 T4779 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@6882a2b1 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1623016b),segFN=segments_1,generation=1}
   [junit4]   2> 2557339 T4779 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2557346 T4895 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard1_replica4] Registered new searcher Searcher@190f9e4a main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2557346 T4797 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2557348 T4797 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2557350 T4797 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2557351 T4797 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2557352 T4797 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2557353 T4797 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2557352 T4770 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard1_replica4
   [junit4]   2> 2557365 T4770 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard1_replica4 address:http://127.0.0.1:50803/i/v collection:awholynewcollection_0 shard:shard1
   [junit4]   2> 2557366 T4770 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2557364 T4779 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 2557368 T4779 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 2557353 T4797 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2557371 T4797 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2557369 T4779 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 2557373 T4779 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 2557374 T4779 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 2557374 T4779 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 2557375 T4779 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 2557381 T4779 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 2557386 T4779 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2557372 T4797 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 2557413 T4797 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2557420 T4779 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 2557442 T4779 oass.SolrIndexSearcher.<init> Opening Searcher@5d7bcff4 main
   [junit4]   2> 2557447 T4779 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2557448 T4779 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2557460 T4797 oass.SolrIndexSearcher.<init> Opening Searcher@1c0055b4 main
   [junit4]   2> 2557466 T4797 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 2557467 T4797 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 2557484 T4896 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard4_replica3] Registered new searcher Searcher@5d7bcff4 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2557489 T4779 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard4_replica3
   [junit4]   2> 2557490 T4779 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard4_replica3 address:http://127.0.0.1:50806/i/v collection:awholynewcollection_0 shard:shard4
   [junit4]   2> 2557491 T4779 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2557492 T4898 oasc.SolrCore.registerSearcher [awholynewcollection_0_shard3_replica1] Registered new searcher Searcher@1c0055b4 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 2557498 T4797 oasc.CoreContainer.registerCore registering core: awholynewcollection_0_shard3_replica1
   [junit4]   2> 2557500 T4797 oasc.ZkController.register Register replica - core:awholynewcollection_0_shard3_replica1 address:http://127.0.0.1:50809/i/v collection:awholynewcollection_0 shard:shard3
   [junit4]   2> 2557500 T4797 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 2557624 T4807 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 2557625 T4807 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2557625 T4807 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/
   [junit4]   2> 2557626 T4807 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard1_replica5 url=http://127.0.0.1:50812/i/v START replicas=[http://127.0.0.1:50809/i/v/awholynewcollection_0_shard1_replica1/, http://127.0.0.1:50803/i/v/awholynewcollection_0_shard1_replica4/, http://127.0.0.1:50806/i/v/awholynewcollection_0_shard1_replica3/, http://127.0.0.1:50799/i/v/awholynewcollection_0_shard1_replica2/] nUpdates=100
   [junit4]   2> 2557628 T4807 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard1_replica5 url=http://127.0.0.1:50812/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2557629 T4807 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2557630 T4807 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2557630 T4807 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/ shard1
   [junit4]   2> 2557630 T4807 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leaders/shard1
   [junit4]   2> ASYNC  NEW_CORE C821 name=awholynewcollection_0_shard1_replica3 org.apache.solr.core.SolrCore@6a9d4194 url=http://127.0.0.1:50806/i/v/awholynewcollection_0_shard1_replica3 node=127.0.0.1:50806_i%2Fv C821_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard1_replica3 props:{state=down, core=awholynewcollection_0_shard1_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2557660 T4783 C821 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=21 
   [junit4]   2> ASYNC  NEW_CORE C822 name=awholynewcollection_0_shard1_replica1 org.apache.solr.core.SolrCore@26d27b42 url=http://127.0.0.1:50809/i/v/awholynewcollection_0_shard1_replica1 node=127.0.0.1:50809_i%2Fv C822_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard1_replica1 props:{state=down, core=awholynewcollection_0_shard1_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2557669 T4793 C822 P50809 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica1] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=6 
   [junit4]   2> ASYNC  NEW_CORE C823 name=awholynewcollection_0_shard1_replica2 org.apache.solr.core.SolrCore@64a2604c url=http://127.0.0.1:50799/i/v/awholynewcollection_0_shard1_replica2 node=127.0.0.1:50799_i%2Fv C823_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard1_replica2 props:{state=down, core=awholynewcollection_0_shard1_replica2, node_name=127.0.0.1:50799_i%2Fv, base_url=http://127.0.0.1:50799/i/v}
   [junit4]   2> 2557681 T4747 C823 P50799 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica2] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=4 
   [junit4]   2> ASYNC  NEW_CORE C824 name=awholynewcollection_0_shard1_replica4 org.apache.solr.core.SolrCore@748bb834 url=http://127.0.0.1:50803/i/v/awholynewcollection_0_shard1_replica4 node=127.0.0.1:50803_i%2Fv C824_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard1_replica4 props:{state=down, core=awholynewcollection_0_shard1_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v}
   [junit4]   2> 2557683 T4897 C824 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=4 
   [junit4]   2> 2557726 T4782 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 2557727 T4782 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2557728 T4782 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/
   [junit4]   2> 2557728 T4782 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard5_replica3 url=http://127.0.0.1:50806/i/v START replicas=[http://127.0.0.1:50809/i/v/awholynewcollection_0_shard5_replica1/, http://127.0.0.1:50799/i/v/awholynewcollection_0_shard5_replica2/, http://127.0.0.1:50803/i/v/awholynewcollection_0_shard5_replica4/, http://127.0.0.1:50812/i/v/awholynewcollection_0_shard5_replica5/] nUpdates=100
   [junit4]   2> 2557732 T4782 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard5_replica3 url=http://127.0.0.1:50806/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2557748 T4782 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2557748 T4782 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2557749 T4782 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/ shard5
   [junit4]   2> 2557750 T4782 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leaders/shard5
   [junit4]   2> 2557754 T4766 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 2557755 T4766 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2557756 T4766 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/
   [junit4]   2> 2557756 T4766 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard3_replica4 url=http://127.0.0.1:50803/i/v START replicas=[http://127.0.0.1:50809/i/v/awholynewcollection_0_shard3_replica1/, http://127.0.0.1:50799/i/v/awholynewcollection_0_shard3_replica2/, http://127.0.0.1:50806/i/v/awholynewcollection_0_shard3_replica3/, http://127.0.0.1:50812/i/v/awholynewcollection_0_shard3_replica5/] nUpdates=100
   [junit4]   2> 2557771 T4766 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard3_replica4 url=http://127.0.0.1:50803/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2557780 T4766 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2557781 T4766 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2557781 T4766 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/ shard3
   [junit4]   2> 2557782 T4766 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leaders/shard3
   [junit4]   2> ASYNC  NEW_CORE C825 name=awholynewcollection_0_shard5_replica1 org.apache.solr.core.SolrCore@35e3126 url=http://127.0.0.1:50809/i/v/awholynewcollection_0_shard5_replica1 node=127.0.0.1:50809_i%2Fv C825_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica1 props:{state=down, core=awholynewcollection_0_shard5_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2557778 T4794 C825 P50809 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica1] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=16 
   [junit4]   2> ASYNC  NEW_CORE C826 name=awholynewcollection_0_shard5_replica5 org.apache.solr.core.SolrCore@47d776f1 url=http://127.0.0.1:50812/i/v/awholynewcollection_0_shard5_replica5 node=127.0.0.1:50812_i%2Fv C826_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica5 props:{state=down, core=awholynewcollection_0_shard5_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2557778 T4808 C826 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=20 
   [junit4]   2> ASYNC  NEW_CORE C827 name=awholynewcollection_0_shard3_replica3 org.apache.solr.core.SolrCore@4180afac url=http://127.0.0.1:50806/i/v/awholynewcollection_0_shard3_replica3 node=127.0.0.1:50806_i%2Fv C827_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica3 props:{state=down, core=awholynewcollection_0_shard3_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2557806 T4781 C827 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=2 
   [junit4]   2> ASYNC  NEW_CORE C828 name=awholynewcollection_0_shard5_replica2 org.apache.solr.core.SolrCore@1a44092b url=http://127.0.0.1:50799/i/v/awholynewcollection_0_shard5_replica2 node=127.0.0.1:50799_i%2Fv C828_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica2 props:{state=down, core=awholynewcollection_0_shard5_replica2, node_name=127.0.0.1:50799_i%2Fv, base_url=http://127.0.0.1:50799/i/v}
   [junit4]   2> 2557817 T4751 C828 P50799 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica2] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=3 
   [junit4]   2> ASYNC  NEW_CORE C829 name=awholynewcollection_0_shard3_replica1 org.apache.solr.core.SolrCore@27a200a0 url=http://127.0.0.1:50809/i/v/awholynewcollection_0_shard3_replica1 node=127.0.0.1:50809_i%2Fv C829_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica1 props:{state=down, core=awholynewcollection_0_shard3_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2557837 T4793 C829 P50809 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica1] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=2 
   [junit4]   2> ASYNC  NEW_CORE C830 name=awholynewcollection_0_shard3_replica2 org.apache.solr.core.SolrCore@16cc90bc url=http://127.0.0.1:50799/i/v/awholynewcollection_0_shard3_replica2 node=127.0.0.1:50799_i%2Fv C830_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica2 props:{state=down, core=awholynewcollection_0_shard3_replica2, node_name=127.0.0.1:50799_i%2Fv, base_url=http://127.0.0.1:50799/i/v}
   [junit4]   2> 2557863 T4751 C830 P50799 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica2] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=4 
   [junit4]   2> ASYNC  NEW_CORE C831 name=awholynewcollection_0_shard3_replica5 org.apache.solr.core.SolrCore@5aad7202 url=http://127.0.0.1:50812/i/v/awholynewcollection_0_shard3_replica5 node=127.0.0.1:50812_i%2Fv C831_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica5 props:{state=down, core=awholynewcollection_0_shard3_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2557869 T4859 C831 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=6 
   [junit4]   2> ASYNC  NEW_CORE C832 name=awholynewcollection_0_shard5_replica4 org.apache.solr.core.SolrCore@587b86 url=http://127.0.0.1:50803/i/v/awholynewcollection_0_shard5_replica4 node=127.0.0.1:50803_i%2Fv C832_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica4 props:{state=down, core=awholynewcollection_0_shard5_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v}
   [junit4]   2> 2557870 T4897 C832 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=2 
   [junit4]   2> 2557905 T4749 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 2557906 T4749 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2557907 T4749 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/
   [junit4]   2> 2557907 T4749 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard4_replica2 url=http://127.0.0.1:50799/i/v START replicas=[http://127.0.0.1:50809/i/v/awholynewcollection_0_shard4_replica1/, http://127.0.0.1:50806/i/v/awholynewcollection_0_shard4_replica3/, http://127.0.0.1:50803/i/v/awholynewcollection_0_shard4_replica4/, http://127.0.0.1:50812/i/v/awholynewcollection_0_shard4_replica5/] nUpdates=100
   [junit4]   2> 2557911 T4749 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard4_replica2 url=http://127.0.0.1:50799/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2557912 T4749 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2557913 T4749 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2557913 T4749 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/ shard4
   [junit4]   2> 2557914 T4749 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leaders/shard4
   [junit4]   2> ASYNC  NEW_CORE C833 name=awholynewcollection_0_shard4_replica3 org.apache.solr.core.SolrCore@adeb916 url=http://127.0.0.1:50806/i/v/awholynewcollection_0_shard4_replica3 node=127.0.0.1:50806_i%2Fv C833_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica3 props:{state=down, core=awholynewcollection_0_shard4_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2557956 T4783 C833 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard4_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=22 
   [junit4]   2> ASYNC  NEW_CORE C834 name=awholynewcollection_0_shard4_replica1 org.apache.solr.core.SolrCore@50d23b4b url=http://127.0.0.1:50809/i/v/awholynewcollection_0_shard4_replica1 node=127.0.0.1:50809_i%2Fv C834_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica1 props:{state=down, core=awholynewcollection_0_shard4_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2557969 T4794 C834 P50809 oasc.SolrCore.execute [awholynewcollection_0_shard4_replica1] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=13 
   [junit4]   2> ASYNC  NEW_CORE C835 name=awholynewcollection_0_shard4_replica5 org.apache.solr.core.SolrCore@57b95f5d url=http://127.0.0.1:50812/i/v/awholynewcollection_0_shard4_replica5 node=127.0.0.1:50812_i%2Fv C835_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica5 props:{state=down, core=awholynewcollection_0_shard4_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2557972 T4808 C835 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard4_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=12 
   [junit4]   2> ASYNC  NEW_CORE C836 name=awholynewcollection_0_shard4_replica4 org.apache.solr.core.SolrCore@430112a4 url=http://127.0.0.1:50803/i/v/awholynewcollection_0_shard4_replica4 node=127.0.0.1:50803_i%2Fv C836_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica4 props:{state=down, core=awholynewcollection_0_shard4_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v}
   [junit4]   2> 2557991 T4897 C836 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard4_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=3 
   [junit4]   2> 2558365 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558368 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node10",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2558380 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2558399 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node9",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2558411 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2558423 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2558483 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2558484 T4831 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2558484 T4761 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2558485 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2558485 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2558486 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2558486 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2558486 T4833 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2558496 T4770 oasc.ZkController.register We are http://127.0.0.1:50803/i/v/awholynewcollection_0_shard1_replica4/ and leader is http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/
   [junit4]   2> 2558497 T4770 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard1_replica4 baseURL=http://127.0.0.1:50803/i/v
   [junit4]   2> 2558498 T4770 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard1_replica4
   [junit4]   2> 2558498 T4770 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558501 T4915 C824 P50803 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard1_replica4 recoveringAfterStartup=true
   [junit4]   2> 2558501 T4915 C824 P50803 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558502 T4915 C824 P50803 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica4 state=recovering
   [junit4]   2> 2558501 T4770 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558503 T4811 oasc.ZkController.register We are http://127.0.0.1:50812/i/v/awholynewcollection_0_shard3_replica5/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/
   [junit4]   2> 2558504 T4811 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard3_replica5 baseURL=http://127.0.0.1:50812/i/v
   [junit4]   2> 2558504 T4811 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard3_replica5
   [junit4]   2> 2558505 T4811 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558507 T4915 C824 P50803 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558507 T4916 C831 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard3_replica5 recoveringAfterStartup=true
   [junit4]   2> 2558508 T4916 C831 P50812 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558508 T4916 C831 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica5 state=recovering
   [junit4]   2> 2558511 T4784 oasc.ZkController.register We are http://127.0.0.1:50806/i/v/awholynewcollection_0_shard3_replica3/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/
   [junit4]   2> 2558512 T4784 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard3_replica3 baseURL=http://127.0.0.1:50806/i/v
   [junit4]   2> 2558512 T4784 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard3_replica3
   [junit4]   2> 2558513 T4784 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558516 T4765 oasc.ZkController.register We are http://127.0.0.1:50803/i/v/awholynewcollection_0_shard5_replica4/ and leader is http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/
   [junit4]   2> 2558517 T4765 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard5_replica4 baseURL=http://127.0.0.1:50803/i/v
   [junit4]   2> 2558517 T4770 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard1&name=awholynewcollection_0_shard1_replica4&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32377 
   [junit4]   2> 2558517 T4798 oasc.ZkController.register We are http://127.0.0.1:50809/i/v/awholynewcollection_0_shard4_replica1/ and leader is http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/
   [junit4]   2> 2558519 T4798 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard4_replica1 baseURL=http://127.0.0.1:50809/i/v
   [junit4]   2> 2558520 T4798 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard4_replica1
   [junit4]   2> 2558520 T4798 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558517 T4765 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard5_replica4
   [junit4]   2> 2558521 T4765 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558520 T4811 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558527 T4917 C827 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard3_replica3 recoveringAfterStartup=true
   [junit4]   2> 2558528 T4917 C827 P50806 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558520 T4782 oasc.ZkController.register We are http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/ and leader is http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/
   [junit4]   2> 2558529 T4782 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 2558530 T4782 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica3 state=active
   [junit4]   2> 2558519 T4795 oasc.ZkController.register We are http://127.0.0.1:50809/i/v/awholynewcollection_0_shard1_replica1/ and leader is http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/
   [junit4]   2> 2558519 T4780 oasc.ZkController.register We are http://127.0.0.1:50806/i/v/awholynewcollection_0_shard1_replica3/ and leader is http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/
   [junit4]   2> 2558532 T4780 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard1_replica3 baseURL=http://127.0.0.1:50806/i/v
   [junit4]   2> 2558532 T4780 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard1_replica3
   [junit4]   2> 2558533 T4780 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558531 T4795 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard1_replica1 baseURL=http://127.0.0.1:50809/i/v
   [junit4]   2> 2558534 T4795 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard1_replica1
   [junit4]   2> 2558535 T4795 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558528 T4917 C827 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica3 state=recovering
   [junit4]   2> 2558536 T4921 C822 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard1_replica1 recoveringAfterStartup=true
   [junit4]   2> 2558536 T4921 C822 P50809 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558537 T4921 C822 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica1 state=recovering
   [junit4]   2> 2558526 T4919 C834 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=true
   [junit4]   2> 2558538 T4919 C834 P50809 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558538 T4919 C834 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovering
   [junit4]   2> 2558526 T4918 C832 P50803 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard5_replica4 recoveringAfterStartup=true
   [junit4]   2> 2558539 T4916 C831 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558539 T4765 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558539 T4918 C832 P50803 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558541 T4918 C832 P50803 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica4 state=recovering
   [junit4]   2> 2558545 T4766 oasc.ZkController.register We are http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/
   [junit4]   2> 2558546 T4766 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 2558547 T4766 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica4 state=active
   [junit4]   2> 2558547 T4807 oasc.ZkController.register We are http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/ and leader is http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/
   [junit4]   2> 2558548 T4807 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 2558548 T4807 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica5 state=active
   [junit4]   2> 2558538 T4784 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558534 T4920 C821 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard1_replica3 recoveringAfterStartup=true
   [junit4]   2> 2558573 T4920 C821 P50806 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558573 T4920 C821 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica3 state=recovering
   [junit4]   2> 2558574 T4797 oasc.ZkController.register We are http://127.0.0.1:50809/i/v/awholynewcollection_0_shard3_replica1/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/
   [junit4]   2> 2558575 T4797 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard3_replica1 baseURL=http://127.0.0.1:50809/i/v
   [junit4]   2> 2558575 T4797 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard3_replica1
   [junit4]   2> 2558576 T4797 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558571 T4798 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558571 T4811 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard3&name=awholynewcollection_0_shard3_replica5&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32350 
   [junit4]   2> 2558579 T4779 oasc.ZkController.register We are http://127.0.0.1:50806/i/v/awholynewcollection_0_shard4_replica3/ and leader is http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/
   [junit4]   2> 2558579 T4779 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard4_replica3 baseURL=http://127.0.0.1:50806/i/v
   [junit4]   2> 2558580 T4779 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard4_replica3
   [junit4]   2> 2558581 T4779 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558570 T4919 C834 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558582 T4923 C833 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=true
   [junit4]   2> 2558582 T4923 C833 P50806 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558583 T4923 C833 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovering
   [junit4]   2> 2558570 T4921 C822 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558569 T4859 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node2, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558594 T4782 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558595 T4920 C821 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558568 T4768 oasc.ZkController.register We are http://127.0.0.1:50803/i/v/awholynewcollection_0_shard4_replica4/ and leader is http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/
   [junit4]   2> 2558603 T4768 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard4_replica4 baseURL=http://127.0.0.1:50803/i/v
   [junit4]   2> 2558604 T4768 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard4_replica4
   [junit4]   2> 2558604 T4768 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558568 T4918 C832 P50803 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558568 T4765 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard5&name=awholynewcollection_0_shard5_replica4&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32266 
   [junit4]   2> 2558606 T4924 C836 P50803 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica4 recoveringAfterStartup=true
   [junit4]   2> 2558595 T4798 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard4&name=awholynewcollection_0_shard4_replica1&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32366 
   [junit4]   2> 2558579 T4922 C829 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard3_replica1 recoveringAfterStartup=true
   [junit4]   2> 2558631 T4922 C829 P50809 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558642 T4922 C829 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica1 state=recovering
   [junit4]   2> 2558579 T4784 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard3&name=awholynewcollection_0_shard3_replica3&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32371 
   [junit4]   2> 2558578 T4807 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558577 T4809 oasc.ZkController.register We are http://127.0.0.1:50812/i/v/awholynewcollection_0_shard4_replica5/ and leader is http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/
   [junit4]   2> 2558648 T4809 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard4_replica5 baseURL=http://127.0.0.1:50812/i/v
   [junit4]   2> 2558649 T4809 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard4_replica5
   [junit4]   2> 2558649 T4809 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558575 T4766 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558656 T4922 C829 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558574 T4917 C827 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558574 T4897 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node15, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558656 T4783 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node24, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558655 T4807 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard1&name=awholynewcollection_0_shard1_replica5&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32517 
   [junit4]   2> 2558652 T4925 C835 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=true
   [junit4]   2> 2558671 T4925 C835 P50812 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558671 T4925 C835 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovering
   [junit4]   2> 2558645 T4780 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558642 T4810 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node3, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558635 T4808 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node1, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558628 T4782 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard5&name=awholynewcollection_0_shard5_replica3&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32335 
   [junit4]   2> 2558627 T4923 C833 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558685 T4766 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard3&name=awholynewcollection_0_shard3_replica4&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32472 
   [junit4]   2> 2558627 T4797 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558626 T4751 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node16, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558625 T4924 C836 P50803 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558698 T4770 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node14, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558697 T4780 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard1&name=awholynewcollection_0_shard1_replica3&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32560 
   [junit4]   2> 2558697 T4925 C835 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558672 T4809 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558701 T4779 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558704 T4768 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558706 T4797 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard3&name=awholynewcollection_0_shard3_replica1&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32519 
   [junit4]   2> 2558700 T4924 C836 P50803 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica4 state=recovering
   [junit4]   2> 2558714 T4809 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard4&name=awholynewcollection_0_shard4_replica5&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32452 
   [junit4]   2> 2558715 T4779 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard4&name=awholynewcollection_0_shard4_replica3&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32463 
   [junit4]   2> 2558716 T4747 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node18, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558727 T4795 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558741 T4768 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard4&name=awholynewcollection_0_shard4_replica4&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32491 
   [junit4]   2> 2558741 T4765 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node11, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558742 T4924 C836 P50803 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558764 T4926 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node19, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558767 T4795 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard1&name=awholynewcollection_0_shard1_replica1&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32634 
   [junit4]   2> 2558768 T4796 oasc.ZkController.register We are http://127.0.0.1:50809/i/v/awholynewcollection_0_shard5_replica1/ and leader is http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/
   [junit4]   2> 2558769 T4796 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard5_replica1 baseURL=http://127.0.0.1:50809/i/v
   [junit4]   2> 2558769 T4796 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard5_replica1
   [junit4]   2> 2558770 T4796 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558772 T4927 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node20, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558773 T4928 C825 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard5_replica1 recoveringAfterStartup=true
   [junit4]   2> 2558774 T4928 C825 P50809 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558774 T4928 C825 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica1 state=recovering
   [junit4]   2> 2558777 T4796 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558780 T4928 C825 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558791 T4796 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard5&name=awholynewcollection_0_shard5_replica1&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32513 
   [junit4]   2> 2558820 T4784 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node21, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558823 T4855 oasc.ZkController.register We are http://127.0.0.1:50812/i/v/awholynewcollection_0_shard5_replica5/ and leader is http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/
   [junit4]   2> 2558824 T4855 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard5_replica5 baseURL=http://127.0.0.1:50812/i/v
   [junit4]   2> 2558824 T4855 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard5_replica5
   [junit4]   2> 2558825 T4855 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558826 T4929 C826 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard5_replica5 recoveringAfterStartup=true
   [junit4]   2> 2558827 T4929 C826 P50812 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558828 T4929 C826 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica5 state=recovering
   [junit4]   2> 2558830 T4855 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558832 T4929 C826 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558844 T4855 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard5&name=awholynewcollection_0_shard5_replica5&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32534 
   [junit4]   2> 2558863 T4782 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node25, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558926 T4750 oasc.ZkController.register We are http://127.0.0.1:50799/i/v/awholynewcollection_0_shard5_replica2/ and leader is http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/
   [junit4]   2> 2558927 T4750 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard5_replica2 baseURL=http://127.0.0.1:50799/i/v
   [junit4]   2> 2558927 T4750 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard5_replica2
   [junit4]   2> 2558928 T4750 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558926 T4749 oasc.ZkController.register We are http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/ and leader is http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/
   [junit4]   2> 2558929 T4749 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 2558929 T4930 C828 P50799 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard5_replica2 recoveringAfterStartup=true
   [junit4]   2> 2558930 T4930 C828 P50799 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558930 T4930 C828 P50799 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica2 state=recovering
   [junit4]   2> 2558930 T4748 oasc.ZkController.register We are http://127.0.0.1:50799/i/v/awholynewcollection_0_shard3_replica2/ and leader is http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/
   [junit4]   2> 2558931 T4748 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard3_replica2 baseURL=http://127.0.0.1:50799/i/v
   [junit4]   2> 2558931 T4748 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard3_replica2
   [junit4]   2> 2558932 T4748 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558930 T4749 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica2 state=active
   [junit4]   2> 2558929 T4752 oasc.ZkController.register We are http://127.0.0.1:50799/i/v/awholynewcollection_0_shard1_replica2/ and leader is http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/
   [junit4]   2> 2558933 T4752 oasc.ZkController.register No LogReplay needed for core=awholynewcollection_0_shard1_replica2 baseURL=http://127.0.0.1:50799/i/v
   [junit4]   2> 2558934 T4752 oasc.ZkController.checkRecovery Core needs to recover:awholynewcollection_0_shard1_replica2
   [junit4]   2> 2558935 T4752 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2558933 T4931 C830 P50799 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard3_replica2 recoveringAfterStartup=true
   [junit4]   2> 2558936 T4931 C830 P50799 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558936 T4931 C830 P50799 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica2 state=recovering
   [junit4]   2> 2558932 T4930 C828 P50799 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558932 T4750 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558936 T4932 C823 P50799 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard1_replica2 recoveringAfterStartup=true
   [junit4]   2> 2558937 T4932 C823 P50799 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 2558938 T4932 C823 P50799 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica2 state=recovering
   [junit4]   2> 2558956 T4931 C830 P50799 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558961 T4932 C823 P50799 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2558991 T4780 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node23, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558993 T4750 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard5&name=awholynewcollection_0_shard5_replica2&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32702 
   [junit4]   2> 2558996 T4749 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2558998 T4766 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node12, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2558999 T4749 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard4&name=awholynewcollection_0_shard4_replica2&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32762 
   [junit4]   2> 2559002 T4752 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2559003 T4807 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node5, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 2559004 T4752 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard1&name=awholynewcollection_0_shard1_replica2&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32862 
   [junit4]   2> 2559007 T4748 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2559009 T4748 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={numShards=5&shard=shard3&name=awholynewcollection_0_shard3_replica2&action=CREATE&collection=awholynewcollection_0&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=32810 
   [junit4]   2> 2559010 T4758 oasc.OverseerCollectionProcessor.createCollection Finished create command on all shards for collection: awholynewcollection_0
   [junit4]   2> 2559013 T4890 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node9, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2559013 T4890 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node9&state=recovering&nodeName=127.0.0.1:50812_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard2_replica4&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2005 
   [junit4]   2> 2559020 T4816 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work/qnr-0000000002 state: SyncConnected type NodeDataChanged
   [junit4]   2> 2559024 T4758 oasc.OverseerCollectionProcessor.run Overseer Collection Processor: Message id:/overseer/collection-queue-work/qn-0000000002 complete, response:{success={null={responseHeader={status=0,QTime=30629},core=awholynewcollection_0_shard2_replica1},null={responseHeader={status=0,QTime=30630},core=awholynewcollection_0_shard2_replica4},null={responseHeader={status=0,QTime=30706},core=awholynewcollection_0_shard2_replica3},null={responseHeader={status=0,QTime=30720},core=awholynewcollection_0_shard2_replica5},null={responseHeader={status=0,QTime=30923},core=awholynewcollection_0_shard2_replica2},null={responseHeader={status=0,QTime=32377},core=awholynewcollection_0_shard1_replica4},null={responseHeader={status=0,QTime=32350},core=awholynewcollection_0_shard3_replica5},null={responseHeader={status=0,QTime=32266},core=awholynewcollection_0_shard5_replica4},null={responseHeader={status=0,QTime=32366},core=awholynewcollection_0_shard4_replica1},null={responseHeader={status=0,QTime=32371},core=awholynewcollection_0_shard3_replica3},null={responseHeader={status=0,QTime=32517},core=awholynewcollection_0_shard1_replica5},null={responseHeader={status=0,QTime=32335},core=awholynewcollection_0_shard5_replica3},null={responseHeader={status=0,QTime=32472},core=awholynewcollection_0_shard3_replica4},null={responseHeader={status=0,QTime=32560},core=awholynewcollection_0_shard1_replica3},null={responseHeader={status=0,QTime=32519},core=awholynewcollection_0_shard3_replica1},null={responseHeader={status=0,QTime=32452},core=awholynewcollection_0_shard4_replica5},null={responseHeader={status=0,QTime=32463},core=awholynewcollection_0_shard4_replica3},null={responseHeader={status=0,QTime=32491},core=awholynewcollection_0_shard4_replica4},null={responseHeader={status=0,QTime=32634},core=awholynewcollection_0_shard1_replica1},null={responseHeader={status=0,QTime=32513},core=awholynewcollection_0_shard5_replica1},null={responseHeader={status=0,QTime=32534},core=awholynewcollection_0_shard5_replica5},null={responseHeader={status=0,QTime=32702},core=awholynewcollection_0_shard5_replica2},null={responseHeader={status=0,QTime=32762},core=awholynewcollection_0_shard4_replica2},null={responseHeader={status=0,QTime=32862},core=awholynewcollection_0_shard1_replica2},null={responseHeader={status=0,QTime=32810},core=awholynewcollection_0_shard3_replica2}}}
   [junit4]   2> 2559029 T4758 oasc.OverseerCollectionProcessor.run Overseer Collection Processor: Get the message id:/overseer/collection-queue-work/qn-0000000004 message:{
   [junit4]   2> 	  "operation":"createcollection",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "maxShardsPerNode":"6",
   [junit4]   2> 	  "createNodeSet":null,
   [junit4]   2> 	  "name":"awholynewcollection_0",
   [junit4]   2> 	  "replicationFactor":"5"}
   [junit4]   2> 2559030 T4758 oasc.SolrException.log ERROR Collection createcollection of createcollection failed:org.apache.solr.common.SolrException: collection already exists: awholynewcollection_0
   [junit4]   2> 		at org.apache.solr.cloud.OverseerCollectionProcessor.createCollection(OverseerCollectionProcessor.java:738)
   [junit4]   2> 		at org.apache.solr.cloud.OverseerCollectionProcessor.processMessage(OverseerCollectionProcessor.java:175)
   [junit4]   2> 		at org.apache.solr.cloud.OverseerCollectionProcessor.run(OverseerCollectionProcessor.java:127)
   [junit4]   2> 		at java.lang.Thread.run(Thread.java:724)
   [junit4]   2> 	
   [junit4]   2> 2559034 T4812 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/collections params={numShards=5&maxShardsPerNode=6&name=awholynewcollection_0&replicationFactor=5&action=CREATE&wt=javabin&version=2} status=0 QTime=32921 
   [junit4]   2> 2559041 T4893 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node8, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2559042 T4893 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node8&state=recovering&nodeName=127.0.0.1:50806_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard2_replica4&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2004 
   [junit4]   2> 2559061 T4774 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work/qnr-0000000004 state: SyncConnected type NodeDataChanged
   [junit4]   2> 2559072 T4767 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node6, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2559073 T4858 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node7, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2559075 T4858 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node7&state=recovering&nodeName=127.0.0.1:50799_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard2_replica4&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2078 
   [junit4]   2> 2559073 T4767 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node6&state=recovering&nodeName=127.0.0.1:50809_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard2_replica4&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2086 
   [junit4]   2> 2559097 T4756 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work state: SyncConnected type NodeChildrenChanged
   [junit4]   2> 2559106 T4769 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/collections params={numShards=5&maxShardsPerNode=6&name=awholynewcollection_0&replicationFactor=5&action=CREATE&wt=javabin&version=2} status=400 QTime=2922 
   [junit4]   2> 2559111 T4758 oasc.OverseerCollectionProcessor.run Overseer Collection Processor: Message id:/overseer/collection-queue-work/qn-0000000004 complete, response:{Operation createcollection caused exception:=org.apache.solr.common.SolrException: collection already exists: awholynewcollection_0,exception={msg=collection already exists: awholynewcollection_0,rspCode=400}}
   [junit4]   2> 2559125 T4736 oazs.NIOServerCnxn.doIO WARN caught end of stream exception EndOfStreamException: Unable to read additional data from client sessionid 0x13fdce320e6000d, likely client has closed socket
   [junit4]   2> 	at org.apache.zookeeper.server.NIOServerCnxn.doIO(NIOServerCnxn.java:220)
   [junit4]   2> 	at org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:208)
   [junit4]   2> 	at java.lang.Thread.run(Thread.java:724)
   [junit4]   2> 
   [junit4]   2> 2559152 T4734 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50799
   [junit4]   2> 2559153 T4734 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=777758190
   [junit4]   2> 2560384 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2560388 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2560402 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node15",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2560415 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node22",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2560437 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node14",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2560453 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2560466 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2560478 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2560491 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node24",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2560514 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node13",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2560528 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2560542 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2560556 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node11",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2560575 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2560590 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node19",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2560603 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node21",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2560616 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node25",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2560631 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node23",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2560651 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node17",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2560671 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node12",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2560684 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2560701 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2560705 T4757 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2> 2560721 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2560723 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2560725 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2560726 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2560727 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2560750 T4765 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node11, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2560752 T4765 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node11&state=recovering&nodeName=127.0.0.1:50809_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard3_replica4&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2011 
   [junit4]   2> 2560825 T4784 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node21, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2560827 T4784 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node21&state=recovering&nodeName=127.0.0.1:50809_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard5_replica3&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2007 
   [junit4]   2> 2560866 T4782 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node25, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2560867 T4782 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node25&state=recovering&nodeName=127.0.0.1:50812_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard5_replica3&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2005 
   [junit4]   2> 2560993 T4780 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node23, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2560995 T4780 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node23&state=recovering&nodeName=127.0.0.1:50799_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard5_replica3&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2004 
   [junit4]   2> 2561001 T4766 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node12, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2561001 T4766 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node12&state=recovering&nodeName=127.0.0.1:50799_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard3_replica4&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2003 
   [junit4]   2> 2561008 T4807 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node5, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 2561008 T4807 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node5&state=recovering&nodeName=127.0.0.1:50799_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard1_replica5&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2005 
   [junit4]   2>  C820_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica5 props:{state=recovering, core=awholynewcollection_0_shard2_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2561016 T4888 C820 P50812 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/ core=awholynewcollection_0_shard2_replica5 - recoveringAfterStartup=true
   [junit4]   2> 2561016 T4888 C820 P50812 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard2_replica5 url=http://127.0.0.1:50812/i/v START replicas=[http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/] nUpdates=100
   [junit4]   2> 2561017 T4888 C820 P50812 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2561017 T4888 C820 P50812 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard2_replica5
   [junit4]   2> 2561017 T4888 C820 P50812 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard2_replica5
   [junit4]   2> 2561017 T4888 C820 P50812 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard2_replica5
   [junit4]   2> 2561018 T4888 C820 P50812 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2561018 T4888 C820 P50812 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/. core=awholynewcollection_0_shard2_replica5
   [junit4]   2> 2561018 T4888 C820 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> ASYNC  NEW_CORE C837 name=awholynewcollection_0_shard2_replica4 org.apache.solr.core.SolrCore@7df4b809 url=http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4 node=127.0.0.1:50803_i%2Fv C837_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica4 props:{state=active, core=awholynewcollection_0_shard2_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v, leader=true}
   [junit4]   2> 2561036 T4768 C837 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2561040 T4858 C837 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2561043 T4858 C837 P50803 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33b1b80c lockFactory=org.apache.lucene.store.NativeFSLockFactory@70f08a18),segFN=segments_1,generation=1}
   [junit4]   2>  C817_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica3 props:{state=recovering, core=awholynewcollection_0_shard2_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2561048 T4887 C817 P50806 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/ core=awholynewcollection_0_shard2_replica3 - recoveringAfterStartup=true
   [junit4]   2> 2561048 T4887 C817 P50806 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard2_replica3 url=http://127.0.0.1:50806/i/v START replicas=[http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/] nUpdates=100
   [junit4]   2> 2561049 T4887 C817 P50806 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2561049 T4887 C817 P50806 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard2_replica3
   [junit4]   2> 2561050 T4887 C817 P50806 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard2_replica3
   [junit4]   2> 2561050 T4887 C817 P50806 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard2_replica3
   [junit4]   2> 2561050 T4887 C817 P50806 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2561051 T4887 C817 P50806 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/. core=awholynewcollection_0_shard2_replica3
   [junit4]   2> 2561051 T4887 C817 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2561056 T4858 C837 P50803 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2561060 T4858 C837 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2561062 T4858 C837 P50803 oass.SolrIndexSearcher.<init> Opening Searcher@454144a1 realtime
   [junit4]   2> 2561062 T4858 C837 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2561063 T4858 C837 P50803 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 23
   [junit4]   2> 2561070 T4888 C820 P50812 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2561071 T4888 C820 P50812 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2561077 T4890 C837 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2561080 T4767 C837 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2561081 T4767 C837 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2561082 T4889 C819 P50799 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/ core=awholynewcollection_0_shard2_replica2 - recoveringAfterStartup=true
   [junit4]   2> 2561082 T4889 C819 P50799 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard2_replica2 url=http://127.0.0.1:50799/i/v START replicas=[http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/] nUpdates=100
   [junit4]   2>  C818_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica1 props:{state=recovering, core=awholynewcollection_0_shard2_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2561083 T4886 C818 P50809 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/ core=awholynewcollection_0_shard2_replica1 - recoveringAfterStartup=true
   [junit4]   2> 2561084 T4886 C818 P50809 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard2_replica1 url=http://127.0.0.1:50809/i/v START replicas=[http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/] nUpdates=100
   [junit4]   2> 2561085 T4889 C819 P50799 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2561085 T4889 C819 P50799 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard2_replica2
   [junit4]   2> 2561086 T4889 C819 P50799 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard2_replica2
   [junit4]   2> 2561086 T4889 C819 P50799 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard2_replica2
   [junit4]   2> 2561086 T4889 C819 P50799 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2561087 T4889 C819 P50799 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/. core=awholynewcollection_0_shard2_replica2
   [junit4]   2> 2561087 T4889 C819 P50799 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2561107 T4890 C837 P50803 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2561108 T4890 C837 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=29 
   [junit4]   2> 2561109 T4886 C818 P50809 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2561110 T4886 C818 P50809 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard2_replica1
   [junit4]   2> 2561110 T4886 C818 P50809 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard2_replica1
   [junit4]   2> 2561110 T4886 C818 P50809 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard2_replica1
   [junit4]   2> 2561111 T4886 C818 P50809 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2561111 T4886 C818 P50809 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard2_replica4/. core=awholynewcollection_0_shard2_replica1
   [junit4]   2> 2561111 T4886 C818 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2561119 T4888 C820 P50812 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard2_replica5
   [junit4]   2> 2561120 T4888 C820 P50812 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard2_replica5
   [junit4]   2> 2561120 T4888 C820 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica5 state=active
   [junit4]   2> 2561128 T4767 C837 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2561129 T4767 C837 P50803 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 49
   [junit4]   2> 2561139 T4893 C837 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2561141 T4887 C817 P50806 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2561161 T4769 C837 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2561170 T4887 C817 P50806 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2561150 T4858 C837 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2561171 T4858 C837 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2561172 T4858 C837 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2561173 T4858 C837 P50803 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 23
   [junit4]   2> 2561174 T4765 C837 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2561175 T4765 C837 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2561175 T4889 C819 P50799 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2561179 T4889 C819 P50799 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2561179 T4767 C837 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=3 
   [junit4]   2> 2561179 T4888 C820 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard2_replica5
   [junit4]   2> 2561181 T4765 C837 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2561189 T4765 C837 P50803 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 28
   [junit4]   2> 2561188 T4766 C837 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=0 
   [junit4]   2> 2561191 T4886 C818 P50809 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2561191 T4886 C818 P50809 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2561191 T4889 C819 P50799 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard2_replica2
   [junit4]   2> 2561194 T4889 C819 P50799 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard2_replica2
   [junit4]   2> 2561194 T4889 C819 P50799 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica2 state=active
   [junit4]   2> 2561191 T4887 C817 P50806 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard2_replica3
   [junit4]   2> 2561194 T4893 C837 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica4] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=0 
   [junit4]   2> 2561196 T4887 C817 P50806 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard2_replica3
   [junit4]   2> 2561197 T4886 C818 P50809 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard2_replica1
   [junit4]   2> 2561201 T4886 C818 P50809 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard2_replica1
   [junit4]   2> 2561201 T4886 C818 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica1 state=active
   [junit4]   2> 2561202 T4887 C817 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica3 state=active
   [junit4]   2> 2561204 T4886 C818 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard2_replica1
   [junit4]   2> 2561206 T4889 C819 P50799 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard2_replica2
   [junit4]   2> 2561208 T4887 C817 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard2_replica3
   [junit4]   2> 2561607 T4859 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node2, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561613 T4859 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node2&state=recovering&nodeName=127.0.0.1:50803_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard1_replica5&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3044 
   [junit4]   2> 2561674 T4783 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node24, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561675 T4783 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node24&state=recovering&nodeName=127.0.0.1:50803_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard5_replica3&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3019 
   [junit4]   2> 2561675 T4897 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node15, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561678 T4897 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node15&state=recovering&nodeName=127.0.0.1:50812_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard3_replica4&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3104 
   [junit4]   2> 2561679 T4810 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node3, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561682 T4810 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node3&state=recovering&nodeName=127.0.0.1:50806_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard1_replica5&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3040 
   [junit4]   2> 2561689 T4808 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node1, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561691 T4808 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node1&state=recovering&nodeName=127.0.0.1:50809_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard1_replica5&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3056 
   [junit4]   2> 2561707 T4770 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node14, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561707 T4751 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node16, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561709 T4751 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node16&state=recovering&nodeName=127.0.0.1:50809_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard4_replica2&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3083 
   [junit4]   2> 2561708 T4770 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node14&state=recovering&nodeName=127.0.0.1:50806_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard3_replica4&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3010 
   [junit4]   2> 2561722 T4747 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node18, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561724 T4747 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node18&state=recovering&nodeName=127.0.0.1:50806_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard4_replica2&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3008 
   [junit4]   2> 2561770 T4926 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node19, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561775 T4926 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node19&state=recovering&nodeName=127.0.0.1:50803_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard4_replica2&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3011 
   [junit4]   2> 2561776 T4927 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node20, state: recovering, checkLive: true, onlyIfLeader: true for: 3 seconds.
   [junit4]   2> 2561776 T4927 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node20&state=recovering&nodeName=127.0.0.1:50812_i%252Fv&action=PREPRECOVERY&checkLive=true&core=awholynewcollection_0_shard4_replica2&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=3004 
   [junit4]   2> 2562177 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica2 state=down
   [junit4]   2> 2562180 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica2 state=down
   [junit4]   2> 2562182 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica2 state=down
   [junit4]   2> 2562184 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica2 state=down
   [junit4]   2> 2562187 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica2 state=down
   [junit4]   2> 2562189 T4734 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node5core=awholynewcollection_0_shard1_replica2
   [junit4]   2> 2562383 T4757 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2562386 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node9",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2562394 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2562402 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2562418 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2562428 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node17",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2562437 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2562446 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2562455 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node12",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2562463 T4757 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node23",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2562471 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2562472 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2562472 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2562481 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2562483 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2>  C829_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica1 props:{state=recovering, core=awholynewcollection_0_shard3_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2562757 T4922 C829 P50809 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/ core=awholynewcollection_0_shard3_replica1 - recoveringAfterStartup=true
   [junit4]   2> 2562757 T4922 C829 P50809 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard3_replica1 url=http://127.0.0.1:50809/i/v START replicas=[http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/] nUpdates=100
   [junit4]   2> 2562758 T4922 C829 P50809 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2562759 T4922 C829 P50809 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard3_replica1
   [junit4]   2> 2562759 T4922 C829 P50809 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard3_replica1
   [junit4]   2> 2562760 T4922 C829 P50809 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard3_replica1
   [junit4]   2> 2562760 T4922 C829 P50809 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2562760 T4922 C829 P50809 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/. core=awholynewcollection_0_shard3_replica1
   [junit4]   2> 2562761 T4922 C829 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> ASYNC  NEW_CORE C838 name=awholynewcollection_0_shard3_replica4 org.apache.solr.core.SolrCore@5475b209 url=http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4 node=127.0.0.1:50803_i%2Fv C838_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica4 props:{state=active, core=awholynewcollection_0_shard3_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v, leader=true}
   [junit4]   2> 2562776 T4766 C838 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2562791 T4768 C838 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2562796 T4768 C838 P50803 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@403564ca lockFactory=org.apache.lucene.store.NativeFSLockFactory@49e455dc),segFN=segments_1,generation=1}
   [junit4]   2> 2562799 T4768 C838 P50803 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2562801 T4768 C838 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2562807 T4768 C838 P50803 oass.SolrIndexSearcher.<init> Opening Searcher@686275e2 realtime
   [junit4]   2> 2562807 T4768 C838 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2562808 T4768 C838 P50803 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 17
   [junit4]   2> 2562810 T4922 C829 P50809 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2562810 T4922 C829 P50809 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2562819 T4769 C838 P50803 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2562821 T4769 C838 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=4 
   [junit4]   2> 2562823 T4922 C829 P50809 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard3_replica1
   [junit4]   2> 2562824 T4922 C829 P50809 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard3_replica1
   [junit4]   2> 2562824 T4922 C829 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica1 state=active
   [junit4]   2>  C825_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica1 props:{state=recovering, core=awholynewcollection_0_shard5_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2562831 T4928 C825 P50809 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/ core=awholynewcollection_0_shard5_replica1 - recoveringAfterStartup=true
   [junit4]   2> 2562832 T4928 C825 P50809 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard5_replica1 url=http://127.0.0.1:50809/i/v START replicas=[http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/] nUpdates=100
   [junit4]   2> 2562833 T4928 C825 P50809 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2562833 T4928 C825 P50809 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard5_replica1
   [junit4]   2> 2562833 T4928 C825 P50809 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard5_replica1
   [junit4]   2> 2562834 T4928 C825 P50809 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard5_replica1
   [junit4]   2> 2562834 T4928 C825 P50809 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2562834 T4928 C825 P50809 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/. core=awholynewcollection_0_shard5_replica1
   [junit4]   2> 2562834 T4928 C825 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2562846 T4922 C829 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard3_replica1
   [junit4]   2> ASYNC  NEW_CORE C839 name=awholynewcollection_0_shard5_replica3 org.apache.solr.core.SolrCore@3e4f8b4b url=http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3 node=127.0.0.1:50806_i%2Fv C839_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica3 props:{state=active, core=awholynewcollection_0_shard5_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v, leader=true}
   [junit4]   2> 2562862 T4784 C839 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2562864 T4779 C839 P50806 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2562867 T4779 C839 P50806 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2d9bb8 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1bf31534),segFN=segments_1,generation=1}
   [junit4]   2> 2562868 T4779 C839 P50806 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2562869 T4779 C839 P50806 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2562870 T4779 C839 P50806 oass.SolrIndexSearcher.<init> Opening Searcher@4e42508d realtime
   [junit4]   2> 2562870 T4779 C839 P50806 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2562871 T4779 C839 P50806 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 8
   [junit4]   2>  C826_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica5 props:{state=recovering, core=awholynewcollection_0_shard5_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2562870 T4929 C826 P50812 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/ core=awholynewcollection_0_shard5_replica5 - recoveringAfterStartup=true
   [junit4]   2> 2562872 T4929 C826 P50812 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard5_replica5 url=http://127.0.0.1:50812/i/v START replicas=[http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/] nUpdates=100
   [junit4]   2> 2562873 T4928 C825 P50809 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2562873 T4928 C825 P50809 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2562873 T4929 C826 P50812 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2562874 T4929 C826 P50812 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard5_replica5
   [junit4]   2> 2562875 T4929 C826 P50812 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard5_replica5
   [junit4]   2> 2562875 T4929 C826 P50812 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard5_replica5
   [junit4]   2> 2562876 T4929 C826 P50812 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2562876 T4929 C826 P50812 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/. core=awholynewcollection_0_shard5_replica5
   [junit4]   2> 2562877 T4929 C826 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2562890 T4782 C839 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2562893 T4782 C839 P50806 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2562894 T4782 C839 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=2 
   [junit4]   2> 2562895 T4928 C825 P50809 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard5_replica1
   [junit4]   2> 2562897 T4928 C825 P50809 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard5_replica1
   [junit4]   2> 2562897 T4928 C825 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica1 state=active
   [junit4]   2> 2562897 T4784 C839 P50806 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2562898 T4784 C839 P50806 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2562899 T4784 C839 P50806 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2562899 T4928 C825 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard5_replica1
   [junit4]   2> 2562899 T4784 C839 P50806 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 2
   [junit4]   2> 2562902 T4929 C826 P50812 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2562903 T4929 C826 P50812 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2562905 T4783 C839 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=0 
   [junit4]   2> 2562906 T4929 C826 P50812 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard5_replica5
   [junit4]   2> 2562907 T4929 C826 P50812 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard5_replica5
   [junit4]   2> 2562907 T4929 C826 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica5 state=active
   [junit4]   2> 2562910 T4929 C826 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard5_replica5
   [junit4]   2> 2562997 T4930 C828 P50799 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/ core=awholynewcollection_0_shard5_replica2 - recoveringAfterStartup=true
   [junit4]   2> 2562998 T4930 C828 P50799 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard5_replica2 url=http://127.0.0.1:50799/i/v START replicas=[http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/] nUpdates=100
   [junit4]   2> 2563000 T4930 C828 P50799 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563001 T4930 C828 P50799 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard5_replica2
   [junit4]   2> 2563001 T4930 C828 P50799 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard5_replica2
   [junit4]   2> 2563002 T4930 C828 P50799 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard5_replica2
   [junit4]   2> 2563002 T4930 C828 P50799 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563002 T4930 C828 P50799 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/. core=awholynewcollection_0_shard5_replica2
   [junit4]   2> 2563003 T4930 C828 P50799 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563004 T4781 C839 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=1 
   [junit4]   2> 2563005 T4931 C830 P50799 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/ core=awholynewcollection_0_shard3_replica2 - recoveringAfterStartup=true
   [junit4]   2> 2563006 T4931 C830 P50799 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard3_replica2 url=http://127.0.0.1:50799/i/v START replicas=[http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/] nUpdates=100
   [junit4]   2> 2563007 T4931 C830 P50799 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563007 T4931 C830 P50799 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard3_replica2
   [junit4]   2> 2563007 T4931 C830 P50799 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard3_replica2
   [junit4]   2> 2563007 T4931 C830 P50799 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard3_replica2
   [junit4]   2> 2563008 T4931 C830 P50799 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563008 T4931 C830 P50799 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/. core=awholynewcollection_0_shard3_replica2
   [junit4]   2> 2563008 T4931 C830 P50799 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563011 T4932 C823 P50799 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/ core=awholynewcollection_0_shard1_replica2 - recoveringAfterStartup=true
   [junit4]   2> 2563011 T4932 C823 P50799 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard1_replica2 url=http://127.0.0.1:50799/i/v START replicas=[http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/] nUpdates=100
   [junit4]   2> 2563017 T4932 C823 P50799 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563018 T4932 C823 P50799 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard1_replica2
   [junit4]   2> 2563018 T4932 C823 P50799 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard1_replica2
   [junit4]   2> 2563019 T4932 C823 P50799 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard1_replica2
   [junit4]   2> 2563019 T4932 C823 P50799 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563019 T4932 C823 P50799 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/. core=awholynewcollection_0_shard1_replica2
   [junit4]   2> 2563020 T4932 C823 P50799 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563020 T4897 C838 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2563032 T4770 C838 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563033 T4770 C838 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2563033 T4770 C838 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563034 T4770 C838 P50803 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 2
   [junit4]   2> 2563036 T4931 C830 P50799 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563036 T4931 C830 P50799 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> ASYNC  NEW_CORE C840 name=awholynewcollection_0_shard1_replica5 org.apache.solr.core.SolrCore@4f6f9093 url=http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5 node=127.0.0.1:50812_i%2Fv C840_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard1_replica5 props:{state=active, core=awholynewcollection_0_shard1_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v, leader=true}
   [junit4]   2> 2563037 T4811 C840 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2563044 T4767 C838 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=1 
   [junit4]   2> 2563048 T4782 C839 P50806 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563050 T4782 C839 P50806 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2563050 T4931 C830 P50799 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard3_replica2
   [junit4]   2> 2563057 T4782 C839 P50806 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563058 T4782 C839 P50806 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 10
   [junit4]   2> 2563060 T4810 C840 P50812 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563061 T4930 C828 P50799 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563062 T4930 C828 P50799 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2563060 T4931 C830 P50799 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard3_replica2
   [junit4]   2> 2563063 T4810 C840 P50812 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191cc979 lockFactory=org.apache.lucene.store.NativeFSLockFactory@193cb3bd),segFN=segments_1,generation=1}
   [junit4]   2> 2563064 T4931 C830 P50799 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica2 state=active
   [junit4]   2> 2563066 T4810 C840 P50812 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 2563068 T4784 C839 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=1 
   [junit4]   2> 2563074 T4810 C840 P50812 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2563078 T4810 C840 P50812 oass.SolrIndexSearcher.<init> Opening Searcher@1a72a1f8 realtime
   [junit4]   2> 2563079 T4810 C840 P50812 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563079 T4810 C840 P50812 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 19
   [junit4]   2> 2563081 T4931 C830 P50799 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard3_replica2
   [junit4]   2> 2563081 T4932 C823 P50799 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563083 T4932 C823 P50799 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2563077 T4930 C828 P50799 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard5_replica2
   [junit4]   2> 2563084 T4930 C828 P50799 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard5_replica2
   [junit4]   2> 2563085 T4930 C828 P50799 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica2 state=active
   [junit4]   2> 2563089 T4807 C840 P50812 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563089 T4807 C840 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=1 
   [junit4]   2> 2563089 T4930 C828 P50799 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard5_replica2
   [junit4]   2> 2563091 T4932 C823 P50799 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard1_replica2
   [junit4]   2> 2563091 T4932 C823 P50799 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard1_replica2
   [junit4]   2> 2563092 T4932 C823 P50799 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica2 state=active
   [junit4]   2> 2563094 T4932 C823 P50799 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard1_replica2
   [junit4]   2> 2563096 T4734 oasc.SolrCore.close [awholynewcollection_0_shard4_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@55063a08
   [junit4]   2> 2563109 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2563110 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2563110 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2563111 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2563112 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard4_replica2] Closing main searcher on request.
   [junit4]   2> 2563113 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2563113 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data/index;done=false>>]
   [junit4]   2> 2563114 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data/index
   [junit4]   2> 2563115 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data;done=false>>]
   [junit4]   2> 2563115 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard4_replica2/data
   [junit4]   2> 2563116 T4734 oasc.SolrCore.close [awholynewcollection_0_shard1_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@64a2604c
   [junit4]   2> 2563137 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2563139 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2563139 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2563140 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2563141 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard1_replica2] Closing main searcher on request.
   [junit4]   2> 2563141 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2563142 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data/index;done=false>>]
   [junit4]   2> 2563142 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data/index
   [junit4]   2> 2563144 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data;done=false>>]
   [junit4]   2> 2563145 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard1_replica2/data
   [junit4]   2> 2563145 T4734 oasc.SolrCore.close [awholynewcollection_0_shard2_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@2dd08a87
   [junit4]   2> 2563158 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2563159 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2563160 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2563161 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2563162 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard2_replica2] Closing main searcher on request.
   [junit4]   2> 2563163 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2563163 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data;done=false>>]
   [junit4]   2> 2563164 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data
   [junit4]   2> 2563165 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data/index;done=false>>]
   [junit4]   2> 2563165 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard2_replica2/data/index
   [junit4]   2> 2563166 T4734 oasc.SolrCore.close [awholynewcollection_0_shard3_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@16cc90bc
   [junit4]   2> 2563179 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2563180 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2563181 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2563182 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2563183 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard3_replica2] Closing main searcher on request.
   [junit4]   2> 2563184 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2563188 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data/index;done=false>>]
   [junit4]   2> 2563190 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data/index
   [junit4]   2> 2563191 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data;done=false>>]
   [junit4]   2> 2563191 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard3_replica2/data
   [junit4]   2> 2563193 T4734 oasc.SolrCore.close [awholynewcollection_0_shard5_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@1a44092b
   [junit4]   2> 2563206 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2563209 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2563210 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2563211 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2563212 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard5_replica2] Closing main searcher on request.
   [junit4]   2> 2563213 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2563213 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data;done=false>>]
   [junit4]   2> 2563214 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data
   [junit4]   2> 2563214 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data/index;done=false>>]
   [junit4]   2> 2563215 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-controljetty-1373800440389//awholynewcollection_0_shard5_replica2/data/index
   [junit4]   2> 2563242 T4757 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90033385638330371-127.0.0.1:50799_i%2Fv-n_0000000000) am no longer a leader.
   [junit4]   2> 2563248 T4756 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 2563260 T4774 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 2563260 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2563262 T4788 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2563262 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2563281 T4802 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2563287 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2563287 T4774 oasc.Overseer.start Overseer (id=90033385638330374-127.0.0.1:50803_i%2Fv-n_0000000001) starting
   [junit4]   2> 2563288 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2563290 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovering, I won't be the leader.
   [junit4]   2> 2563293 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2563293 T4734 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/i/v,null}
   [junit4]   2> 2563296 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2563297 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563306 T4944 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 2563307 T4774 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2563307 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2563312 T4943 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 2563314 T4774 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2563321 T4774 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2563321 T4774 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovering, I won't be the leader.
   [junit4]   2> 2563322 T4774 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2563321 T4943 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2563325 T4774 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2563325 T4774 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node19core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563327 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node11",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2563331 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2563357 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2563361 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovering, I won't be the leader.
   [junit4]   2> 2563362 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2563362 T4734 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50803
   [junit4]   2> 2563364 T4734 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=2042028234
   [junit4]   2> 2563366 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2563366 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563376 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node21",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2563380 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2563392 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2563392 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovering, I won't be the leader.
   [junit4]   2> 2563393 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2563398 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2563399 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563403 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node25",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2563430 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node12",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2563439 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node23",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2563448 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica2",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50799_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50799/i/v"}
   [junit4]   2> 2563483 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2563484 T4943 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 2563492 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node10",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2563493 T4943 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2> 2563508 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node13",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2563508 T4943 oasc.Overseer$ClusterStateUpdater.updateState shard=shard3 is already registered
   [junit4]   2> 2563521 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node19",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2563521 T4943 oasc.Overseer$ClusterStateUpdater.updateState shard=shard4 is already registered
   [junit4]   2> 2563531 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node24",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2563531 T4943 oasc.Overseer$ClusterStateUpdater.updateState shard=shard5 is already registered
   [junit4]   2> 2563540 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2563540 T4943 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2>  C824_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard1_replica4 props:{state=recovering, core=awholynewcollection_0_shard1_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v}
   [junit4]   2> 2563618 T4915 C824 P50803 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/ core=awholynewcollection_0_shard1_replica4 - recoveringAfterStartup=true
   [junit4]   2> 2563619 T4915 C824 P50803 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard1_replica4 url=http://127.0.0.1:50803/i/v START replicas=[http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/] nUpdates=100
   [junit4]   2> 2563620 T4915 C824 P50803 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563620 T4915 C824 P50803 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard1_replica4
   [junit4]   2> 2563620 T4915 C824 P50803 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard1_replica4
   [junit4]   2> 2563621 T4915 C824 P50803 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard1_replica4
   [junit4]   2> 2563621 T4915 C824 P50803 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563621 T4915 C824 P50803 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/. core=awholynewcollection_0_shard1_replica4
   [junit4]   2> 2563621 T4915 C824 P50803 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563625 T4810 C840 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2563641 T4807 C840 P50812 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563642 T4807 C840 P50812 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2563643 T4807 C840 P50812 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563643 T4807 C840 P50812 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 2
   [junit4]   2> 2563645 T4915 C824 P50803 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563645 T4915 C824 P50803 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2563652 T4809 C840 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=2 
   [junit4]   2> 2563655 T4915 C824 P50803 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard1_replica4
   [junit4]   2> 2563655 T4915 C824 P50803 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard1_replica4
   [junit4]   2> 2563655 T4915 C824 P50803 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica4 state=active
   [junit4]   2> 2563658 T4915 C824 P50803 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard1_replica4
   [junit4]   2>  C832_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica4 props:{state=recovering, core=awholynewcollection_0_shard5_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v}
   [junit4]   2> 2563678 T4918 C832 P50803 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/ core=awholynewcollection_0_shard5_replica4 - recoveringAfterStartup=true
   [junit4]   2> 2563679 T4918 C832 P50803 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard5_replica4 url=http://127.0.0.1:50803/i/v START replicas=[http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/] nUpdates=100
   [junit4]   2> 2563680 T4918 C832 P50803 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563680 T4918 C832 P50803 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard5_replica4
   [junit4]   2> 2563680 T4918 C832 P50803 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard5_replica4
   [junit4]   2> 2563681 T4918 C832 P50803 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard5_replica4
   [junit4]   2>  C831_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica5 props:{state=recovering, core=awholynewcollection_0_shard3_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2563681 T4916 C831 P50812 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/ core=awholynewcollection_0_shard3_replica5 - recoveringAfterStartup=true
   [junit4]   2> 2563682 T4916 C831 P50812 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard3_replica5 url=http://127.0.0.1:50812/i/v START replicas=[http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/] nUpdates=100
   [junit4]   2> 2563681 T4918 C832 P50803 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563682 T4918 C832 P50803 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50806/i/v/awholynewcollection_0_shard5_replica3/. core=awholynewcollection_0_shard5_replica4
   [junit4]   2> 2563683 T4918 C832 P50803 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563684 T4783 C839 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2563683 T4916 C831 P50812 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563685 T4916 C831 P50812 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard3_replica5
   [junit4]   2> 2563686 T4916 C831 P50812 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard3_replica5
   [junit4]   2> 2563686 T4916 C831 P50812 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard3_replica5
   [junit4]   2> 2563686 T4916 C831 P50812 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563687 T4916 C831 P50812 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/. core=awholynewcollection_0_shard3_replica5
   [junit4]   2> 2563687 T4916 C831 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2>  C821_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard1_replica3 props:{state=recovering, core=awholynewcollection_0_shard1_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2563693 T4920 C821 P50806 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/ core=awholynewcollection_0_shard1_replica3 - recoveringAfterStartup=true
   [junit4]   2> 2563693 T4920 C821 P50806 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard1_replica3 url=http://127.0.0.1:50806/i/v START replicas=[http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/] nUpdates=100
   [junit4]   2>  C822_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard1_replica1 props:{state=recovering, core=awholynewcollection_0_shard1_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2563694 T4921 C822 P50809 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/ core=awholynewcollection_0_shard1_replica1 - recoveringAfterStartup=true
   [junit4]   2> 2563694 T4921 C822 P50809 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard1_replica1 url=http://127.0.0.1:50809/i/v START replicas=[http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/] nUpdates=100
   [junit4]   2> 2563701 T4920 C821 P50806 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563704 T4920 C821 P50806 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard1_replica3
   [junit4]   2> 2563704 T4920 C821 P50806 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard1_replica3
   [junit4]   2> 2563704 T4920 C821 P50806 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard1_replica3
   [junit4]   2> 2563705 T4920 C821 P50806 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563705 T4920 C821 P50806 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/. core=awholynewcollection_0_shard1_replica3
   [junit4]   2> 2563705 T4920 C821 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563703 T4858 C838 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=1 
   [junit4]   2> 2563703 T4921 C822 P50809 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563711 T4921 C822 P50809 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard1_replica1
   [junit4]   2> 2563711 T4921 C822 P50809 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard1_replica1
   [junit4]   2> 2563711 T4921 C822 P50809 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard1_replica1
   [junit4]   2> 2563712 T4921 C822 P50809 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563712 T4921 C822 P50809 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50812/i/v/awholynewcollection_0_shard1_replica5/. core=awholynewcollection_0_shard1_replica1
   [junit4]   2> 2563712 T4921 C822 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563714 T4855 C840 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2563715 T4855 C840 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2>  C834_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica1 props:{state=recovering, core=awholynewcollection_0_shard4_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2563717 T4919 C834 P50809 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/ core=awholynewcollection_0_shard4_replica1 - recoveringAfterStartup=true
   [junit4]   2> 2563717 T4919 C834 P50809 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard4_replica1 url=http://127.0.0.1:50809/i/v START replicas=[http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/] nUpdates=100
   [junit4]   2>  C827_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica3 props:{state=recovering, core=awholynewcollection_0_shard3_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2563718 T4917 C827 P50806 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/ core=awholynewcollection_0_shard3_replica3 - recoveringAfterStartup=true
   [junit4]   2> 2563718 T4917 C827 P50806 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard3_replica3 url=http://127.0.0.1:50806/i/v START replicas=[http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/] nUpdates=100
   [junit4]   2> 2563719 T4919 C834 P50809 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563719 T4919 C834 P50809 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563720 T4919 C834 P50809 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563720 T4919 C834 P50809 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563720 T4919 C834 P50809 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563720 T4919 C834 P50809 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563721 T4919 C834 P50809 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563722 T4770 C838 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563723 T4770 C838 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2>  C833_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica3 props:{state=recovering, core=awholynewcollection_0_shard4_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2563727 T4923 C833 P50806 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/ core=awholynewcollection_0_shard4_replica3 - recoveringAfterStartup=true
   [junit4]   2> 2563728 T4923 C833 P50806 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard4_replica3 url=http://127.0.0.1:50806/i/v START replicas=[http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/] nUpdates=100
   [junit4]   2> 2563728 T4923 C833 P50806 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563729 T4923 C833 P50806 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563729 T4923 C833 P50806 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563730 T4923 C833 P50806 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563730 T4923 C833 P50806 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563731 T4923 C833 P50806 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563731 T4923 C833 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563730 T4917 C827 P50806 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563732 T4917 C827 P50806 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard3_replica3
   [junit4]   2> 2563732 T4917 C827 P50806 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard3_replica3
   [junit4]   2> 2563733 T4917 C827 P50806 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard3_replica3
   [junit4]   2> 2563733 T4917 C827 P50806 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563733 T4917 C827 P50806 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50803/i/v/awholynewcollection_0_shard3_replica4/. core=awholynewcollection_0_shard3_replica3
   [junit4]   2> 2563734 T4917 C827 P50806 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563740 T4770 C838 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563741 T4770 C838 P50803 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 20
   [junit4]   2> 2563749 T4890 C838 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2563750 T4781 C839 P50806 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563769 T4781 C839 P50806 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2563765 T4811 C840 P50812 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563771 T4811 C840 P50812 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2563772 T4811 C840 P50812 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563772 T4811 C840 P50812 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 7
   [junit4]   2> 2563762 T4916 C831 P50812 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563774 T4916 C831 P50812 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2563772 T4810 C840 P50812 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563775 T4810 C840 P50812 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2563770 T4765 C838 P50803 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 2563776 T4765 C838 P50803 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 2563777 T4765 C838 P50803 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563777 T4765 C838 P50803 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 7
   [junit4]   2>  C836_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica4 props:{state=recovering, core=awholynewcollection_0_shard4_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v}
   [junit4]   2> 2563779 T4924 C836 P50803 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/ core=awholynewcollection_0_shard4_replica4 - recoveringAfterStartup=true
   [junit4]   2> 2563780 T4924 C836 P50803 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard4_replica4 url=http://127.0.0.1:50803/i/v START replicas=[http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/] nUpdates=100
   [junit4]   2> 2563780 T4917 C827 P50806 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563781 T4917 C827 P50806 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2563776 T4919 C834 P50809 oasc.SolrException.log ERROR Error while trying to recover:org.apache.solr.client.solrj.SolrServerException: Server refused connection at: http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:428)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)
   [junit4]   2> 		at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:117)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.commitOnLeader(RecoveryStrategy.java:179)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.replicate(RecoveryStrategy.java:130)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:405)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	Caused by: org.apache.http.conn.HttpHostConnectException: Connection to http://127.0.0.1:50799 refused
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:190)
   [junit4]   2> 		at org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:294)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:645)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:480)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:906)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:805)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:365)
   [junit4]   2> 		... 6 more
   [junit4]   2> 	Caused by: java.net.ConnectException: Connection refused
   [junit4]   2> 		at java.net.PlainSocketImpl.socketConnect(Native Method)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182)
   [junit4]   2> 		at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
   [junit4]   2> 		at java.net.Socket.connect(Socket.java:579)
   [junit4]   2> 		at org.apache.http.conn.scheme.PlainSocketFactory.connectSocket(PlainSocketFactory.java:127)
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:180)
   [junit4]   2> 		... 13 more
   [junit4]   2> 	
   [junit4]   2> 2563782 T4919 C834 P50809 oasu.UpdateLog.dropBufferedUpdates Dropping buffered updates FSUpdateLog{state=BUFFERING, tlog=null}
   [junit4]   2> 2563783 T4919 C834 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563783 T4919 C834 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563783 T4919 C834 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563783 T4919 C834 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2563781 T4924 C836 P50803 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563785 T4924 C836 P50803 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563785 T4924 C836 P50803 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563786 T4924 C836 P50803 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563786 T4924 C836 P50803 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563787 T4924 C836 P50803 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563787 T4924 C836 P50803 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563795 T4893 C838 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=0 
   [junit4]   2>  C835_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica5 props:{state=recovering, core=awholynewcollection_0_shard4_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2563780 T4925 C835 P50812 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/ core=awholynewcollection_0_shard4_replica5 - recoveringAfterStartup=true
   [junit4]   2> 2563797 T4925 C835 P50812 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard4_replica5 url=http://127.0.0.1:50812/i/v START replicas=[http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/] nUpdates=100
   [junit4]   2> 2563798 T4925 C835 P50812 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 2563798 T4925 C835 P50812 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563799 T4925 C835 P50812 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563799 T4925 C835 P50812 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563800 T4925 C835 P50812 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 2563800 T4925 C835 P50812 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2/. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563801 T4925 C835 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563807 T4924 C836 P50803 oasc.SolrException.log ERROR Error while trying to recover:org.apache.solr.client.solrj.SolrServerException: Server refused connection at: http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:428)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)
   [junit4]   2> 		at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:117)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.commitOnLeader(RecoveryStrategy.java:179)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.replicate(RecoveryStrategy.java:130)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:405)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	Caused by: org.apache.http.conn.HttpHostConnectException: Connection to http://127.0.0.1:50799 refused
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:190)
   [junit4]   2> 		at org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:294)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:645)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:480)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:906)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:805)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:365)
   [junit4]   2> 		... 6 more
   [junit4]   2> 	Caused by: java.net.ConnectException: Connection refused
   [junit4]   2> 		at java.net.PlainSocketImpl.socketConnect(Native Method)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182)
   [junit4]   2> 		at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
   [junit4]   2> 		at java.net.Socket.connect(Socket.java:579)
   [junit4]   2> 		at org.apache.http.conn.scheme.PlainSocketFactory.connectSocket(PlainSocketFactory.java:127)
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:180)
   [junit4]   2> 		... 13 more
   [junit4]   2> 	
   [junit4]   2> 2563808 T4924 C836 P50803 oasu.UpdateLog.dropBufferedUpdates Dropping buffered updates FSUpdateLog{state=BUFFERING, tlog=null}
   [junit4]   2> 2563808 T4924 C836 P50803 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563808 T4924 C836 P50803 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563809 T4924 C836 P50803 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563809 T4924 C836 P50803 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica4 state=recovery_failed
   [junit4]   2> 2563778 T4921 C822 P50809 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563819 T4921 C822 P50809 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2563777 T4781 C839 P50806 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563821 T4781 C839 P50806 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 71
   [junit4]   2> 2563777 T4923 C833 P50806 oasc.SolrException.log ERROR Error while trying to recover:org.apache.solr.client.solrj.SolrServerException: Server refused connection at: http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:428)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)
   [junit4]   2> 		at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:117)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.commitOnLeader(RecoveryStrategy.java:179)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.replicate(RecoveryStrategy.java:130)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:405)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	Caused by: org.apache.http.conn.HttpHostConnectException: Connection to http://127.0.0.1:50799 refused
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:190)
   [junit4]   2> 		at org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:294)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:645)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:480)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:906)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:805)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:365)
   [junit4]   2> 		... 6 more
   [junit4]   2> 	Caused by: java.net.ConnectException: Connection refused
   [junit4]   2> 		at java.net.PlainSocketImpl.socketConnect(Native Method)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182)
   [junit4]   2> 		at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
   [junit4]   2> 		at java.net.Socket.connect(Socket.java:579)
   [junit4]   2> 		at org.apache.http.conn.scheme.PlainSocketFactory.connectSocket(PlainSocketFactory.java:127)
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:180)
   [junit4]   2> 		... 13 more
   [junit4]   2> 	
   [junit4]   2> 2563824 T4923 C833 P50806 oasu.UpdateLog.dropBufferedUpdates Dropping buffered updates FSUpdateLog{state=BUFFERING, tlog=null}
   [junit4]   2> 2563825 T4923 C833 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563825 T4923 C833 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563825 T4923 C833 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563826 T4923 C833 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2563824 T4807 C840 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=1 
   [junit4]   2> 2563807 T4919 C834 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563829 T4919 C834 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563799 T4917 C827 P50806 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard3_replica3
   [junit4]   2> 2563829 T4917 C827 P50806 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard3_replica3
   [junit4]   2> 2563786 T4810 C840 P50812 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 2563830 T4955 C834 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2563831 T4917 C827 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica3 state=active
   [junit4]   2> 2563785 T4768 C838 P50803 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica4] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=0 
   [junit4]   2> 2563831 T4923 C833 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563834 T4923 C833 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563830 T4810 C840 P50812 oasup.LogUpdateProcessor.finish [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 65
   [junit4]   2> 2563829 T4921 C822 P50809 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard1_replica1
   [junit4]   2> 2563836 T4921 C822 P50809 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard1_replica1
   [junit4]   2> 2563837 T4921 C822 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica1 state=active
   [junit4]   2> 2563839 T4956 C833 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2563836 T4917 C827 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard3_replica3
   [junit4]   2> 2563844 T4920 C821 P50806 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563852 T4920 C821 P50806 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2563845 T4921 C822 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard1_replica1
   [junit4]   2> 2563833 T4916 C831 P50812 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard3_replica5
   [junit4]   2> 2563855 T4916 C831 P50812 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard3_replica5
   [junit4]   2> 2563855 T4916 C831 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica5 state=active
   [junit4]   2> 2563851 T4918 C832 P50803 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 2563858 T4918 C832 P50803 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 2563849 T4924 C836 P50803 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node19core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563860 T4924 C836 P50803 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563861 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2563862 T4957 C836 P50803 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica4 recoveringAfterStartup=false
   [junit4]   2> 2563865 T4916 C831 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard3_replica5
   [junit4]   2> 2563845 T4925 C835 P50812 oasc.SolrException.log ERROR Error while trying to recover:org.apache.solr.client.solrj.SolrServerException: Server refused connection at: http://127.0.0.1:50799/i/v/awholynewcollection_0_shard4_replica2
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:428)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)
   [junit4]   2> 		at org.apache.solr.client.solrj.request.AbstractUpdateRequest.process(AbstractUpdateRequest.java:117)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.commitOnLeader(RecoveryStrategy.java:179)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.replicate(RecoveryStrategy.java:130)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:405)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	Caused by: org.apache.http.conn.HttpHostConnectException: Connection to http://127.0.0.1:50799 refused
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:190)
   [junit4]   2> 		at org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:294)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:645)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:480)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:906)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:805)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:365)
   [junit4]   2> 		... 6 more
   [junit4]   2> 	Caused by: java.net.ConnectException: Connection refused
   [junit4]   2> 		at java.net.PlainSocketImpl.socketConnect(Native Method)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182)
   [junit4]   2> 		at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
   [junit4]   2> 		at java.net.Socket.connect(Socket.java:579)
   [junit4]   2> 		at org.apache.http.conn.scheme.PlainSocketFactory.connectSocket(PlainSocketFactory.java:127)
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:180)
   [junit4]   2> 		... 13 more
   [junit4]   2> 	
   [junit4]   2> 2563862 T4780 C839 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica3] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=0 
   [junit4]   2> 2563857 T4809 C840 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard1_replica5] webapp=/i/v path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=0 
   [junit4]   2> 2563875 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2563876 T4920 C821 P50806 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard1_replica3
   [junit4]   2> 2563876 T4920 C821 P50806 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard1_replica3
   [junit4]   2> 2563877 T4920 C821 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica3 state=active
   [junit4]   2> 2563873 T4925 C835 P50812 oasu.UpdateLog.dropBufferedUpdates Dropping buffered updates FSUpdateLog{state=BUFFERING, tlog=null}
   [junit4]   2> 2563878 T4925 C835 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563878 T4925 C835 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563876 T4918 C832 P50803 oasc.RecoveryStrategy.replay No replay needed. core=awholynewcollection_0_shard5_replica4
   [junit4]   2> 2563880 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2563883 T4918 C832 P50803 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=awholynewcollection_0_shard5_replica4
   [junit4]   2> 2563883 T4918 C832 P50803 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica4 state=active
   [junit4]   2> 2563879 T4925 C835 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563886 T4925 C835 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2563886 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2563887 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2563888 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2563886 T4920 C821 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard1_replica3
   [junit4]   2> 2563899 T4918 C832 P50803 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard5_replica4
   [junit4]   2> 2563899 T4925 C835 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563900 T4925 C835 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563902 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2563902 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2563905 T4958 C835 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2563906 T4958 C835 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovering
   [junit4]   2> 2563909 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2563912 T4958 C835 P50812 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 2563931 T4816 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 2563931 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2563932 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2563932 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2563931 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2563937 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2563937 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2563938 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2563948 T4958 C835 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.client.solrj.SolrServerException: Server refused connection at: http://127.0.0.1:50799/i/v
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:428)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.sendPrepRecoveryCmd(RecoveryStrategy.java:198)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:342)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	Caused by: org.apache.http.conn.HttpHostConnectException: Connection to http://127.0.0.1:50799 refused
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:190)
   [junit4]   2> 		at org.apache.http.impl.conn.ManagedClientConnectionImpl.open(ManagedClientConnectionImpl.java:294)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.tryConnect(DefaultRequestDirector.java:645)
   [junit4]   2> 		at org.apache.http.impl.client.DefaultRequestDirector.execute(DefaultRequestDirector.java:480)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:906)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:805)
   [junit4]   2> 		at org.apache.http.impl.client.AbstractHttpClient.execute(AbstractHttpClient.java:784)
   [junit4]   2> 		at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:365)
   [junit4]   2> 		... 4 more
   [junit4]   2> 	Caused by: java.net.ConnectException: Connection refused
   [junit4]   2> 		at java.net.PlainSocketImpl.socketConnect(Native Method)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:339)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:200)
   [junit4]   2> 		at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:182)
   [junit4]   2> 		at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
   [junit4]   2> 		at java.net.Socket.connect(Socket.java:579)
   [junit4]   2> 		at org.apache.http.conn.scheme.PlainSocketFactory.connectSocket(PlainSocketFactory.java:127)
   [junit4]   2> 		at org.apache.http.impl.conn.DefaultClientConnectionOperator.openConnection(DefaultClientConnectionOperator.java:180)
   [junit4]   2> 		... 11 more
   [junit4]   2> 	
   [junit4]   2> 2563949 T4958 C835 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2563949 T4958 C835 P50812 oasc.RecoveryStrategy.doRecovery Wait 2.0 seconds before trying to recover again (1)
   [junit4]   2> 2563950 T4774 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2563956 T4774 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2563957 T4774 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2563957 T4774 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2563960 T4774 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2563960 T4774 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node19core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2563962 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2563968 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2563969 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovering, I won't be the leader.
   [junit4]   2> 2563969 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2563970 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2563971 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2564550 T4756 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2564551 T4756 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2564855 T4955 C834 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2564856 T4955 C834 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2564857 T4955 C834 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2564857 T4955 C834 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2564858 T4955 C834 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2564860 T4955 C834 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2564860 T4955 C834 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2564862 T4959 C834 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2564867 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2564870 T4956 C833 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2564870 T4956 C833 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2564870 T4956 C833 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2564871 T4956 C833 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2564871 T4956 C833 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2564874 T4956 C833 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2564875 T4956 C833 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2564882 T4960 C833 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2564884 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2564885 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2564886 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2>  C836_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica4 props:{state=down, core=awholynewcollection_0_shard4_replica4, node_name=127.0.0.1:50803_i%2Fv, base_url=http://127.0.0.1:50803/i/v}
   [junit4]   2> 2564888 T4957 C836 P50803 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica4:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2564889 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2564889 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2564889 T4957 C836 P50803 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2564890 T4957 C836 P50803 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2564891 T4957 C836 P50803 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2564891 T4957 C836 P50803 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica4 state=recovery_failed
   [junit4]   2> 2564891 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2564893 T4957 C836 P50803 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node19core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2564894 T4957 C836 P50803 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2564895 T4961 C836 P50803 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica4 recoveringAfterStartup=false
   [junit4]   2> 2564897 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2564898 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2564898 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2564901 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2564901 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2564903 T4774 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2564908 T4774 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2564908 T4774 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2564908 T4774 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2564911 T4774 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2564911 T4774 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node19core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2564952 T4958 C835 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2564959 T4962 C835 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2564964 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2564971 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2564972 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovering, I won't be the leader.
   [junit4]   2> 2564972 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2564974 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2564975 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2565118 T4943 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2565122 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2565133 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2565141 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node19",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2565149 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2565159 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node14",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2565173 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2565183 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node15",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2565193 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2565215 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2565237 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node24",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2565249 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2565286 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2565295 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2565327 T4943 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node19",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2565412 T4734 oasc.ZkController.publish publishing core=nodes_used_collection_shard2_replica1 state=down
   [junit4]   2> 2565417 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica4 state=down
   [junit4]   2> 2565425 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica4 state=down
   [junit4]   2> 2565429 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica4 state=down
   [junit4]   2> 2565431 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica4 state=down
   [junit4]   2> 2565433 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica4 state=down
   [junit4]   2> 2565892 T4959 C834 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2565893 T4959 C834 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2565893 T4959 C834 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2565894 T4959 C834 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2565894 T4959 C834 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2565896 T4959 C834 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2565897 T4959 C834 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2565898 T4963 C834 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2565902 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2565907 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2565907 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2565908 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2565908 T4960 C833 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2565909 T4960 C833 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2565909 T4960 C833 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2565909 T4960 C833 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2565910 T4960 C833 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2565914 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2565914 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2565915 T4960 C833 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2565916 T4960 C833 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2565917 T4964 C833 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2565921 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2565922 T4961 C836 P50803 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica4:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2565923 T4961 C836 P50803 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2565923 T4961 C836 P50803 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2565923 T4961 C836 P50803 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2565924 T4961 C836 P50803 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica4 state=recovery_failed
   [junit4]   2> 2565927 T4961 C836 P50803 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node19core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2565927 T4961 C836 P50803 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2565928 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2565928 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2565929 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2565929 T4965 C836 P50803 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica4 recoveringAfterStartup=false
   [junit4]   2> 2565930 T4965 C836 P50803 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2565929 T4734 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node19core=awholynewcollection_0_shard4_replica4
   [junit4]   2> 2565935 T4734 oasc.SolrCore.close [nodes_used_collection_shard2_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@22cda955
   [junit4]   2> 2565935 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2565936 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2565937 T4774 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2565949 T4774 oasc.LeaderElector$1.process WARN  org.apache.solr.common.SolrException: Fatal Error, SolrCore not found:awholynewcollection_0_shard4_replica4 in []
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:163)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.access$000(LeaderElector.java:55)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector$1.process(LeaderElector.java:129)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:519)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:495)
   [junit4]   2> 
   [junit4]   2> 2565950 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2565952 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=1,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2565953 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2565954 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2565954 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2565959 T4734 oasc.SolrCore.closeSearcher [nodes_used_collection_shard2_replica1] Closing main searcher on request.
   [junit4]   2> 2565960 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2565961 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data/index;done=false>>]
   [junit4]   2> 2565962 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data/index
   [junit4]   2> 2565963 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data;done=false>>]
   [junit4]   2> 2565963 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//nodes_used_collection_shard2_replica1/data
   [junit4]   2> 2565964 T4734 oasc.SolrCore.close [awholynewcollection_0_shard2_replica4]  CLOSING SolrCore org.apache.solr.core.SolrCore@7df4b809
   [junit4]   2> 2565986 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=4,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2565987 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2565987 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2565988 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2565992 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard2_replica4] Closing main searcher on request.
   [junit4]   2> 2565992 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2565993 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data/index;done=false>>]
   [junit4]   2> 2565994 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data/index
   [junit4]   2> 2565995 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data;done=false>>]
   [junit4]   2> 2565996 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard2_replica4/data
   [junit4]   2> 2565996 T4734 oasc.SolrCore.close [awholynewcollection_0_shard3_replica4]  CLOSING SolrCore org.apache.solr.core.SolrCore@5475b209
   [junit4]   2> 2565998 T4962 C835 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2565999 T4962 C835 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2565999 T4962 C835 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2566000 T4962 C835 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2566001 T4962 C835 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2566003 T4962 C835 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2566004 T4962 C835 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2566005 T4966 C835 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2566011 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=4,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2566013 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2566013 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2566015 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2566015 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2566039 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard3_replica4] Closing main searcher on request.
   [junit4]   2> 2566041 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2566042 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data/index;done=false>>]
   [junit4]   2> 2566043 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data/index
   [junit4]   2> 2566044 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data;done=false>>]
   [junit4]   2> 2566045 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard3_replica4/data
   [junit4]   2> 2566045 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2566046 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2566047 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2566046 T4734 oasc.SolrCore.close [awholynewcollection_0_shard4_replica4]  CLOSING SolrCore org.apache.solr.core.SolrCore@430112a4
   [junit4]   2> 2566059 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2566060 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2566063 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2566064 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2566064 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2566065 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2566066 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard4_replica4] Closing main searcher on request.
   [junit4]   2> 2566067 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2566068 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data;done=false>>]
   [junit4]   2> 2566068 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data
   [junit4]   2> 2566069 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data/index;done=false>>]
   [junit4]   2> 2566070 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard4_replica4/data/index
   [junit4]   2> 2566070 T4734 oasc.SolrCore.close [awholynewcollection_0_shard5_replica4]  CLOSING SolrCore org.apache.solr.core.SolrCore@587b86
   [junit4]   2> 2566091 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2566093 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2566093 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2566094 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2566095 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard5_replica4] Closing main searcher on request.
   [junit4]   2> 2566096 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2566096 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data;done=false>>]
   [junit4]   2> 2566097 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data
   [junit4]   2> 2566098 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data/index;done=false>>]
   [junit4]   2> 2566099 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard5_replica4/data/index
   [junit4]   2> 2566099 T4734 oasc.SolrCore.close [awholynewcollection_0_shard1_replica4]  CLOSING SolrCore org.apache.solr.core.SolrCore@748bb834
   [junit4]   2> 2566233 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2566235 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2566235 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2566236 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2566237 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard1_replica4] Closing main searcher on request.
   [junit4]   2> 2566237 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2566238 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data/index;done=false>>]
   [junit4]   2> 2566239 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data/index
   [junit4]   2> 2566239 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data;done=false>>]
   [junit4]   2> 2566240 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty1-1373800442008//awholynewcollection_0_shard1_replica4/data
   [junit4]   2> 2566242 T4943 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90033385638330374-127.0.0.1:50803_i%2Fv-n_0000000001) am no longer a leader.
   [junit4]   2> 2566264 T4734 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/i/v,null}
   [junit4]   2> 2566317 T4734 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50806
   [junit4]   2> 2566318 T4734 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1444740563
   [junit4]   2> ASYNC  NEW_CORE C841 name=awholynewcollection_0_shard4_replica1 org.apache.solr.core.SolrCore@50d23b4b url=http://127.0.0.1:50809/i/v/awholynewcollection_0_shard4_replica1 node=127.0.0.1:50809_i%2Fv C841_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica1 props:{state=recovering, core=awholynewcollection_0_shard4_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2566930 T4963 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2566930 T4963 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2566931 T4963 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2566931 T4963 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2566931 T4963 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2566933 T4963 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2566934 T4963 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2566935 T4967 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2566939 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> ASYNC  NEW_CORE C842 name=awholynewcollection_0_shard4_replica3 org.apache.solr.core.SolrCore@adeb916 url=http://127.0.0.1:50806/i/v/awholynewcollection_0_shard4_replica3 node=127.0.0.1:50806_i%2Fv C842_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica3 props:{state=recovering, core=awholynewcollection_0_shard4_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2566941 T4964 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2566941 T4964 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2566942 T4964 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2566942 T4964 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2566943 T4964 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2566947 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2566947 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2566947 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2566949 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2566950 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2566952 T4964 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2566953 T4964 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2566954 T4968 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2566959 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2566973 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2566974 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2566975 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2566979 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2566981 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> ASYNC  NEW_CORE C843 name=awholynewcollection_0_shard4_replica5 org.apache.solr.core.SolrCore@57b95f5d url=http://127.0.0.1:50812/i/v/awholynewcollection_0_shard4_replica5 node=127.0.0.1:50812_i%2Fv C843_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica5 props:{state=recovering, core=awholynewcollection_0_shard4_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2567049 T4966 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2567050 T4966 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2567051 T4966 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2567051 T4966 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2567051 T4966 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2567055 T4966 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2567056 T4966 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2567058 T4969 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2567065 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2567073 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2567074 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2567074 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2567078 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2567079 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2567547 T4774 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2567549 T4774 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2567549 T4774 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2567966 T4967 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2567967 T4967 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2567967 T4967 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2567968 T4967 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2567968 T4967 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2567971 T4967 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2567971 T4967 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2567973 T4970 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2567978 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2567984 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2567985 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2567985 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2567990 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2567991 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2567991 T4968 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2567992 T4968 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2567992 T4968 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2567993 T4968 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2567993 T4968 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2567995 T4968 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2567995 T4968 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2567996 T4971 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2568000 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2568005 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2568005 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2568006 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2568008 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2568009 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2568096 T4969 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2568097 T4969 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2568097 T4969 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2568098 T4969 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2568098 T4969 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2568100 T4969 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2568101 T4969 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2568102 T4972 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2568105 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2568148 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2568149 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2568149 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2568153 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2568154 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2568999 T4970 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2569000 T4970 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2569000 T4970 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2569001 T4970 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2569001 T4970 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2569004 T4970 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2569004 T4970 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2569006 T4973 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2569011 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2569020 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2569022 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2569025 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2569027 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2569028 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2569034 T4971 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2569034 T4971 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2569035 T4971 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2569035 T4971 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2569035 T4971 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2569038 T4971 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2569038 T4971 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2569040 T4974 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2569045 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2569051 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2569051 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2569052 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2569054 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2569054 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2569138 T4972 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2569139 T4972 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2569139 T4972 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2569139 T4972 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2569140 T4972 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2569142 T4972 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2569143 T4972 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2569144 T4975 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2569147 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2569152 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2569152 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2569153 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2569161 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2569162 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2570037 T4973 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2570038 T4973 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2570038 T4973 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2570039 T4973 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2570039 T4973 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2570042 T4973 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2570042 T4973 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2570044 T4976 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2570048 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2570054 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2570054 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2570055 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2570057 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2570058 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2570075 T4974 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2570075 T4974 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2570076 T4974 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2570076 T4974 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2570077 T4974 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2570079 T4974 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2570079 T4974 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2570081 T4977 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2570091 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2570096 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2570096 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2570097 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2570099 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2570100 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2570177 T4975 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2570177 T4975 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2570178 T4975 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2570178 T4975 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2570178 T4975 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2570181 T4975 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2570181 T4975 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2570183 T4978 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2570186 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2570191 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2570191 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2570192 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2570193 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2570194 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2571075 T4976 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2571075 T4976 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2571075 T4976 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2571076 T4976 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2571076 T4976 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2571078 T4976 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2571079 T4976 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2571080 T4979 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2571089 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2571094 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2571095 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2571095 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2571097 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2571098 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2571111 T4977 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2571112 T4977 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2571112 T4977 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2571113 T4977 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2571113 T4977 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2571115 T4977 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2571116 T4977 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2571117 T4980 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2571123 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2571130 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2571131 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2571132 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2571135 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2571136 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2571228 T4978 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2571229 T4978 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2571229 T4978 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2571230 T4978 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2571230 T4978 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2571233 T4978 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2571234 T4978 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2571235 T4981 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2571241 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2571247 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2571248 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2571248 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2571250 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2571251 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2572111 T4979 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2572112 T4979 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2572112 T4979 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2572112 T4979 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2572113 T4979 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2572115 T4979 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2572116 T4979 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2572117 T4982 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2572122 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2572127 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2572128 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2572128 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2572131 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2572131 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2572156 T4980 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2572157 T4980 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2572157 T4980 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2572158 T4980 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2572158 T4980 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2572161 T4980 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2572161 T4980 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2572162 T4983 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2572167 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2572172 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2572173 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2572174 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2572176 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2572177 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2572268 T4981 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2572269 T4981 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2572269 T4981 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2572270 T4981 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2572271 T4981 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2572274 T4981 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2572275 T4981 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2572276 T4984 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2572286 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2572291 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2572292 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2572292 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2572295 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2572295 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2573152 T4982 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2573153 T4982 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2573153 T4982 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2573153 T4982 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2573154 T4982 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2573162 T4982 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2573163 T4982 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2573164 T4985 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2573169 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2573175 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2573176 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2573176 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2573178 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2573179 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2573192 T4983 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2573193 T4983 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2573193 T4983 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2573193 T4983 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2573193 T4983 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2573196 T4983 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2573196 T4983 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2573197 T4986 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2573202 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2573206 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2573206 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2573207 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2573209 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2573209 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2573317 T4984 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2573318 T4984 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2573318 T4984 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2573319 T4984 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2573319 T4984 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2573329 T4984 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2573329 T4984 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2573330 T4987 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2573334 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2573338 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2573338 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2573339 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2573340 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2573341 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2574207 T4985 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2574208 T4985 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2574208 T4985 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2574208 T4985 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2574208 T4985 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2574211 T4985 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2574211 T4985 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2574215 T4988 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2574224 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2574227 T4986 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2574228 T4986 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2574228 T4986 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2574228 T4986 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2574229 T4986 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2574229 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2574232 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2574233 T4986 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2574234 T4986 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2574233 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2574235 T4989 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2574237 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2574237 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2574240 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2574245 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2574246 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2574246 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2574248 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2574249 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2574372 T4987 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2574373 T4987 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2574374 T4987 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2574374 T4987 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2574375 T4987 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2574379 T4987 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2574380 T4987 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2574384 T4990 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2574395 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2574409 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2574410 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2574411 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2574415 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2574416 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2575248 T4988 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2575248 T4988 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2575249 T4988 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2575249 T4988 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2575249 T4988 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2575251 T4988 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2575252 T4988 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2575253 T4991 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2575257 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2575260 T4989 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2575261 T4989 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2575261 T4989 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2575261 T4989 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2575262 T4989 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2575264 T4989 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2575265 T4989 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2575266 T4992 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2575267 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2575268 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2575268 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2575270 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2575272 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2575272 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2575277 T4788 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 2575282 T4788 oasc.Overseer.start Overseer (id=90033385638330376-127.0.0.1:50806_i%2Fv-n_0000000002) starting
   [junit4]   2> 2575292 T4994 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 2575294 T4788 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 2575295 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 2575299 T4993 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 2575301 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2575306 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2575307 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2575308 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2575308 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2575312 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2575312 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2575313 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2575328 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node10",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2575337 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node13",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2575348 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node19",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2575365 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node24",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2575376 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2575385 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575407 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575413 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node19",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica4",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50803_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50803/i/v"}
   [junit4]   2> 2575414 T4990 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2575414 T4990 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2575414 T4990 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2575414 T4990 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2575415 T4990 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2575420 T4990 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2575420 T4990 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2575422 T4995 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2575427 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2575435 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2575436 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2575437 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2575439 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2575440 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2575467 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2575486 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575486 T4993 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 2575494 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575494 T4993 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2> 2575502 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node14",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575503 T4993 oasc.Overseer$ClusterStateUpdater.updateState shard=shard3 is already registered
   [junit4]   2> 2575516 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575516 T4993 oasc.Overseer$ClusterStateUpdater.updateState shard=shard4 is already registered
   [junit4]   2> 2575526 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node22",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575527 T4993 oasc.Overseer$ClusterStateUpdater.updateState shard=shard5 is already registered
   [junit4]   2> 2575538 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"nodes_used_collection_shard1_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575538 T4993 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 2575546 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575562 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575592 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2575609 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575629 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575648 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2575666 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575685 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575715 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2575733 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575748 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575762 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2575785 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575800 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575814 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2575837 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575855 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575877 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2575892 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575912 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575928 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2575946 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2575964 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2575991 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2576009 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2576020 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2576055 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2576276 T4991 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2576276 T4991 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2576277 T4991 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2576277 T4991 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2576277 T4991 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2576280 T4991 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2576280 T4991 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2576281 T4996 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2576286 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2576295 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2576295 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2576295 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2576297 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2576297 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2>  C842_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica3 props:{state=recovery_failed, core=awholynewcollection_0_shard4_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2576297 T4992 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2576298 T4992 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2576299 T4992 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2576299 T4992 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2576299 T4992 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2576301 T4992 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2576302 T4992 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2576303 T4997 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2576311 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2576319 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2576319 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2576320 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2576323 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2576324 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2576448 T4995 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2576449 T4995 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2576449 T4995 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2576449 T4995 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2576450 T4995 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2576452 T4995 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2576452 T4995 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2576453 T4998 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2576457 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2576461 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2576462 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2576462 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2576464 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2576465 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2577307 T4996 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2577307 T4996 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2577308 T4996 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2577308 T4996 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2577308 T4996 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2577311 T4996 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2577311 T4996 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2577313 T4999 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2577317 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2577322 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2577323 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2577323 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2577325 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2577325 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2577333 T4997 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2577333 T4997 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2577334 T4997 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2577334 T4997 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2577334 T4997 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2577336 T4997 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2577337 T4997 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2577338 T5000 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2577342 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2577347 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2577348 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2577348 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2577349 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2577350 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2577488 T4998 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2577491 T4998 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2577494 T4998 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2577496 T4998 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2577497 T4998 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2577499 T4998 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2577499 T4998 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2577501 T5001 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2577507 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2577512 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2577513 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2577514 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2577516 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2577516 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2577898 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2577901 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2577917 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2577932 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2577954 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2577972 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2577988 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2578316 T4999 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2578317 T4999 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2578317 T4999 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2578318 T4999 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2578318 T4999 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2578321 T4999 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2578322 T4999 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2578323 T5002 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2578328 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2578334 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2578336 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2578336 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2578340 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2578341 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2578373 T5000 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2578381 T5000 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2578382 T5000 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2578382 T5000 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2578383 T5000 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2578386 T5000 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2578387 T5000 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2578388 T5003 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2578394 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2578400 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2578400 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2578401 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2578403 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2578404 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2578533 T5001 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2578534 T5001 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2578534 T5001 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2578535 T5001 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2578535 T5001 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2578538 T5001 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2578538 T5001 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2578540 T5004 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2578545 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2578551 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2578552 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2578552 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2578555 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2578555 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2579350 T5002 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2579351 T5002 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2579351 T5002 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2579352 T5002 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2579352 T5002 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2579354 T5002 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2579355 T5002 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2579357 T5005 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2579361 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2579366 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2579367 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2579367 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2579369 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2579370 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2579424 T5003 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2579424 T5003 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2579425 T5003 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2579425 T5003 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2579425 T5003 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2579427 T5003 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2579428 T5003 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2579429 T5006 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2579433 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2579437 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2579438 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2579438 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2579440 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2579441 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2579558 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2579562 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2579578 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2579579 T5004 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2579579 T5004 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2579580 T5004 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2579580 T5004 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2579581 T5004 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2579584 T5004 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2579584 T5004 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2579586 T5007 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2579593 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2579600 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2579600 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2579601 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2579610 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2579611 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2579612 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2579635 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2579652 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2579668 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2580390 T5005 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2580391 T5005 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2580391 T5005 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2580391 T5005 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2580392 T5005 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2580394 T5005 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2580395 T5005 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2580396 T5008 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2580402 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2580409 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2580410 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2580411 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2580414 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2580415 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2580464 T5006 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2580464 T5006 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2580464 T5006 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2580465 T5006 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2580465 T5006 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2580468 T5006 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2580469 T5006 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2580470 T5009 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2580475 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2580481 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2580482 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2580482 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2580485 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2580485 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2580617 T5007 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2580618 T5007 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2580619 T5007 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2580619 T5007 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2580620 T5007 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2580623 T5007 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2580623 T5007 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2580625 T5010 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2580630 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2580635 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2580636 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2580637 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2580639 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2580640 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2581274 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2581279 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2581304 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2581316 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2581426 T5008 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2581433 T5008 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2581434 T5008 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2581435 T5008 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2581435 T5008 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2581438 T5008 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2581439 T5008 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2581441 T5011 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2581446 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2581454 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2581456 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2581457 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2581461 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2581462 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2581484 T5009 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2581485 T5009 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2581486 T5009 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2581486 T5009 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2581487 T5009 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2581490 T5009 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2581491 T5009 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2581493 T5012 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2581500 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2581508 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2581509 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2581509 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2581512 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2581513 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2581641 T5010 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2581642 T5010 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2581643 T5010 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2581644 T5010 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2581644 T5010 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2581648 T5010 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2581649 T5010 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2581651 T5013 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2581657 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2581665 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2581667 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2581668 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2581671 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2581673 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2582465 T5011 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2582466 T5011 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2582466 T5011 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2582467 T5011 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2582467 T5011 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2582471 T5011 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2582472 T5011 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2582474 T5014 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2582480 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2582500 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2582502 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2582504 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2582507 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2582509 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2582519 T5012 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2582520 T5012 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2582521 T5012 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2582522 T5012 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2582522 T5012 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2582526 T5012 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2582527 T5012 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2582528 T5015 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2582535 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2582543 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2582544 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2582545 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2582548 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2582549 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2582678 T5013 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2582682 T5013 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2582685 T5013 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2582691 T5013 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2582692 T5013 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2582695 T5013 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2582696 T5013 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2582697 T5016 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2582704 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2582713 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2582714 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2582715 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2582718 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2582720 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2582900 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2582903 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2582936 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2582953 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2582969 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2582997 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2583023 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2583516 T5014 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2583517 T5014 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2583517 T5014 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2583517 T5014 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2583518 T5014 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2583521 T5014 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2583522 T5014 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2583524 T5017 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2583531 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2583539 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2583540 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2583542 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2583545 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2583545 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2583575 T5015 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2583578 T5015 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2583579 T5015 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2583579 T5015 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2583579 T5015 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2583582 T5015 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2583582 T5015 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2583583 T5018 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2583588 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2583593 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2583593 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2583594 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2583596 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2583596 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2583726 T5016 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2583727 T5016 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2583727 T5016 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2583728 T5016 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2583728 T5016 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2583731 T5016 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2583731 T5016 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2583733 T5019 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2583737 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2583742 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2583743 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2583743 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2583747 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2583752 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2584553 T5017 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2584556 T5017 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2584556 T5017 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2584557 T5017 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2584557 T5017 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2584562 T5017 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2584563 T5017 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2584564 T5020 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2584569 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2584575 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2584576 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2584577 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2584580 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2584581 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2584611 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2584611 T5018 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2584612 T5018 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2584612 T5018 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2584613 T5018 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2584613 T5018 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2584613 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2584625 T5018 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2584626 T5018 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2584627 T5021 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2584646 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2584652 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2584653 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2584654 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2584657 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2584658 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2584660 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2584677 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2584696 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2584715 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2584762 T5019 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2584763 T5019 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2584763 T5019 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2584764 T5019 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2584764 T5019 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2584769 T5019 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2584769 T5019 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2584771 T5022 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2584775 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2584785 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2584786 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2584787 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2584789 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2584790 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2585603 T5020 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2585604 T5020 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2585604 T5020 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2585604 T5020 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2585605 T5020 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2585607 T5020 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2585607 T5020 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2585608 T5023 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2585612 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2585618 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2585621 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2585622 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2585626 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2585627 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2585657 T5021 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2585657 T5021 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2585657 T5021 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2585658 T5021 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2585658 T5021 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2585661 T5021 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2585661 T5021 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2585662 T5024 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2585667 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2585672 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2585672 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2585672 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2585674 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2585675 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2585805 T5022 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2585806 T5022 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2585806 T5022 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2585806 T5022 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2585807 T5022 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2585809 T5022 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2585809 T5022 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2585810 T5025 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2585814 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2585824 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2585825 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2585826 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2585828 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2585828 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2586308 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2586311 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2586333 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2586369 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2586391 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2586639 T5023 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2586639 T5023 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2586640 T5023 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2586640 T5023 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2586640 T5023 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2586643 T5023 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2586644 T5023 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2586645 T5026 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2586649 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2586655 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2586662 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2586663 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2586670 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2586674 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2586694 T5024 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2586695 T5024 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2586696 T5024 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2586697 T5024 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2586697 T5024 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2586699 T5024 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2586700 T5024 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2586701 T5027 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2586705 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2586710 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2586711 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2586712 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2586714 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2586715 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2586846 T5025 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2586847 T5025 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2586847 T5025 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2586847 T5025 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2586848 T5025 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2586850 T5025 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2586851 T5025 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2586853 T5028 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2586859 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2586866 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2586868 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2586869 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2586873 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2586875 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2587677 T5026 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2587678 T5026 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2587678 T5026 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2587679 T5026 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2587679 T5026 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2587685 T5026 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2587685 T5026 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2587687 T5029 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2587694 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2587705 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2587707 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2587708 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2587710 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2587711 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2587735 T5027 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2587735 T5027 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2587736 T5027 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2587736 T5027 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2587736 T5027 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2587739 T5027 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2587739 T5027 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2587740 T5030 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2587745 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2587750 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2587751 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2587751 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2587753 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2587754 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2587891 T5028 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2587891 T5028 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2587891 T5028 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2587892 T5028 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2587893 T5028 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2587895 T5028 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2587896 T5028 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2587897 T5031 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2587902 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2587909 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2587910 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2587911 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2587915 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2587916 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2587948 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2587953 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2587985 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2588008 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2588031 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2588061 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2588085 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2588716 T5029 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2588718 T5029 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2588719 T5029 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2588719 T5029 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2588720 T5029 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2588723 T5029 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2588724 T5029 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2588725 T5032 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2588730 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2588736 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2588737 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2588737 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2588740 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2588740 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2588767 T5030 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2588768 T5030 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2588770 T5030 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2588770 T5030 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2588771 T5030 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2588773 T5030 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2588774 T5030 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2588776 T5033 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2588781 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2588789 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2588790 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2588792 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2588794 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2588795 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2588938 T5031 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2588939 T5031 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2588939 T5031 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2588940 T5031 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2588941 T5031 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2588943 T5031 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2588944 T5031 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2588945 T5034 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2588953 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2588966 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2588967 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2588968 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2588971 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2588972 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2589687 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2589691 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2589709 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2589724 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2589759 T5032 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2589760 T5032 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2589760 T5032 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2589760 T5032 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2589761 T5032 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2589763 T5032 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2589763 T5032 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2589765 T5035 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2589769 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2589775 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2589775 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2589776 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2589778 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2589779 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2589809 T5033 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2589810 T5033 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2589813 T5033 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2589814 T5033 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2589816 T5033 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2589818 T5033 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2589819 T5033 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2589820 T5036 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2589824 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2589829 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2589830 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2589830 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2589834 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2589834 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2589975 T5034 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2589976 T5034 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2589976 T5034 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2589977 T5034 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2589977 T5034 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2589984 T5034 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2589984 T5034 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2589986 T5037 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2589993 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2590001 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2590003 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2590004 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2590008 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2590009 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2590791 T5035 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2590792 T5035 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2590792 T5035 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2590792 T5035 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2590793 T5035 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2590795 T5035 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2590796 T5035 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2590797 T5038 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2590801 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2590807 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2590807 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2590808 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2590815 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2590817 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2590854 T5036 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2590854 T5036 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2590855 T5036 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2590855 T5036 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2590855 T5036 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2590858 T5036 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2590858 T5036 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2590859 T5039 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2590863 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2590868 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2590869 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2590869 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2590875 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2590879 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2591021 T5037 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2591021 T5037 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2591022 T5037 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2591022 T5037 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2591022 T5037 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2591025 T5037 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2591025 T5037 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2591029 T5040 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2591034 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2591041 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2591041 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2591042 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2591045 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2591047 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2591278 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2591287 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2591312 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2591333 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2591356 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2591373 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2591396 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2591829 T5038 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2591829 T5038 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2591830 T5038 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2591830 T5038 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2591831 T5038 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2591837 T5038 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2591839 T5038 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2591840 T5041 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2591845 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2591850 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2591851 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2591851 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2591856 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2591857 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2591883 T5039 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2591884 T5039 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2591884 T5039 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2591884 T5039 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2591885 T5039 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2591887 T5039 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2591887 T5039 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2591888 T5042 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2591893 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2591898 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2591898 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2591899 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2591901 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2591902 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2592063 T5040 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2592063 T5040 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2592064 T5040 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2592064 T5040 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2592064 T5040 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2592066 T5040 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2592067 T5040 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2592068 T5043 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2592072 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2592083 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2592084 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2592085 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2592087 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2592088 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2592877 T5041 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2592878 T5041 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2592878 T5041 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2592879 T5041 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2592879 T5041 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2592881 T5041 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2592881 T5041 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2592882 T5044 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2592886 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2592890 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2592891 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2592892 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2592893 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2592894 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2592931 T5042 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2592932 T5042 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2592932 T5042 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2592932 T5042 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2592932 T5042 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2592935 T5042 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2592935 T5042 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2592937 T5045 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2592941 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2592946 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2592947 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2592947 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2592950 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2592950 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2592972 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2592975 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2593001 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2593019 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2593036 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2593054 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2593095 T5043 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2593096 T5043 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2593096 T5043 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2593096 T5043 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2593097 T5043 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2593100 T5043 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2593100 T5043 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2593103 T5046 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2593110 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2593122 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2593123 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2593124 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2593127 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2593128 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2593918 T5044 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2593919 T5044 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2593919 T5044 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2593919 T5044 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2593920 T5044 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2593922 T5044 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2593923 T5044 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2593924 T5047 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2593928 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2593934 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2593935 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2593936 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2593938 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2593939 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2593969 T5045 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2593970 T5045 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2593970 T5045 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2593970 T5045 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2593971 T5045 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2593973 T5045 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2593974 T5045 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2593975 T5048 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2593980 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2593989 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2593991 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2593991 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2593994 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2593995 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2594153 T5046 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2594154 T5046 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2594154 T5046 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2594155 T5046 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2594155 T5046 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2594158 T5046 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2594159 T5046 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2594160 T5049 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2594165 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2594171 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2594172 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2594173 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2594175 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2594176 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2594640 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2594645 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2594666 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2594682 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2594699 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2594957 T5047 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2594958 T5047 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2594958 T5047 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2594958 T5047 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2594959 T5047 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2594961 T5047 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2594961 T5047 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2594963 T5050 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2594966 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2594971 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2594972 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2594973 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2594976 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2594976 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2595007 T5048 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2595007 T5048 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2595011 T5048 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2595011 T5048 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2595012 T5048 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2595015 T5048 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2595015 T5048 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2595017 T5051 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2595023 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2595028 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2595028 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2595029 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2595030 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2595031 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2595206 T5049 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2595206 T5049 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2595207 T5049 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2595207 T5049 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2595207 T5049 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2595210 T5049 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2595210 T5049 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2595211 T5052 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2595216 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2595223 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2595224 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2595226 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2595229 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2595229 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2596002 T5050 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2596002 T5050 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2596003 T5050 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2596003 T5050 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2596003 T5050 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2596006 T5050 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2596006 T5050 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2596010 T5053 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2596018 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2596025 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2596027 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2596027 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2596030 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2596031 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2596064 T5051 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2596065 T5051 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2596065 T5051 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2596065 T5051 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2596066 T5051 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2596068 T5051 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2596068 T5051 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2596070 T5054 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2596073 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2596087 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2596088 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2596089 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2596092 T4788 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2596092 T4788 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2596246 T5052 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2596247 T5052 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2596247 T5052 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2596248 T5052 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2596248 T5052 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2596250 T5052 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2596251 T5052 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2596252 T5055 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2596257 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2596262 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2596263 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2596264 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2596265 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2596266 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2596295 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2596301 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2596317 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2596334 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2596355 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2596369 T4734 oasc.ZkController.publishAndWaitForDownStates WARN Timed out waiting to see all nodes published as DOWN in our cluster state.
   [junit4]   2> 2596370 T4734 oasc.ZkController.publish publishing core=nodes_used_collection_shard1_replica1 state=down
   [junit4]   2> 2596373 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica3 state=down
   [junit4]   2> 2596376 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica3 state=down
   [junit4]   2> 2596379 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica3 state=down
   [junit4]   2> 2596381 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2596383 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica3 state=down
   [junit4]   2> 2596386 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=down
   [junit4]   2> 2596403 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2596426 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard1_replica1",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2596435 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node22",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2596444 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2596453 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node14",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2596462 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2596471 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2597062 T5053 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2597063 T5053 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2597063 T5053 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2597063 T5053 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2597064 T5053 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2597067 T5053 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2597067 T5053 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2597068 T5056 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2597073 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2597085 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2597086 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2597087 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2597093 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2597094 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2597116 T5054 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2597117 T5054 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2597117 T5054 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2597118 T5054 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2597118 T5054 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2597122 T5054 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2597123 T5054 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2597125 T5057 C842 P50806 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica3 recoveringAfterStartup=false
   [junit4]   2> 2597126 T4734 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2597134 T4788 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2597141 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2597143 T4788 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2597144 T4788 oasc.ShardLeaderElectionContext.rejoinLeaderElection Not rejoining election because CoreContainer is shutdown
   [junit4]   2> 2597145 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 2597290 T5055 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2597291 T5055 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2597291 T5055 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2597291 T5055 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2597292 T5055 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2597294 T5055 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2597294 T5055 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2597296 T5058 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2597301 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2597305 T4816 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 2597305 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 2597310 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard3
   [junit4]   2> 2597315 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2597315 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 2597316 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2597316 T4816 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50812/i/v/awholynewcollection_0_shard3_replica5/
   [junit4]   2> 2597316 T4816 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard3_replica5 url=http://127.0.0.1:50812/i/v START replicas=[http://127.0.0.1:50809/i/v/awholynewcollection_0_shard3_replica1/, http://127.0.0.1:50806/i/v/awholynewcollection_0_shard3_replica3/] nUpdates=100
   [junit4]   2> 2597318 T4816 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard3_replica5 url=http://127.0.0.1:50812/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2597319 T4816 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2597319 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2597319 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50812/i/v/awholynewcollection_0_shard3_replica5/ shard3
   [junit4]   2> 2597320 T4816 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leaders/shard3
   [junit4]   2> ASYNC  NEW_CORE C844 name=awholynewcollection_0_shard3_replica1 org.apache.solr.core.SolrCore@27a200a0 url=http://127.0.0.1:50809/i/v/awholynewcollection_0_shard3_replica1 node=127.0.0.1:50809_i%2Fv C844_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica1 props:{state=recovering, core=awholynewcollection_0_shard3_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2597331 T4795 C844 P50809 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica1] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=1 
   [junit4]   2> ASYNC  NEW_CORE C845 name=awholynewcollection_0_shard3_replica3 org.apache.solr.core.SolrCore@4180afac url=http://127.0.0.1:50806/i/v/awholynewcollection_0_shard3_replica3 node=127.0.0.1:50806_i%2Fv C845_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard3_replica3 props:{state=down, core=awholynewcollection_0_shard3_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2597334 T4782 C845 P50806 oasc.SolrCore.execute [awholynewcollection_0_shard3_replica3] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2598075 T4993 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2598078 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2598094 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2598098 T5056 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2598098 T5056 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2598099 T5056 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2598099 T5056 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2598099 T5056 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2598111 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2598113 T5056 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2598114 T5056 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2598117 T5061 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2598126 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 2598130 T4802 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 2598131 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 2598139 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 2598148 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2>  C842_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica3 props:{state=down, core=awholynewcollection_0_shard4_replica3, node_name=127.0.0.1:50806_i%2Fv, base_url=http://127.0.0.1:50806/i/v}
   [junit4]   2> 2598154 T5057 C842 P50806 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica3:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2598155 T5057 C842 P50806 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2598155 T5057 C842 P50806 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2598155 T5057 C842 P50806 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2598156 T5057 C842 P50806 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica3 state=recovery_failed
   [junit4]   2> 2598156 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2598157 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 2598157 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2598157 T4802 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50809/i/v/nodes_used_collection_shard2_replica2/
   [junit4]   2> 2598158 T4802 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 2598158 T4802 oasc.SyncStrategy.syncToMe http://127.0.0.1:50809/i/v/nodes_used_collection_shard2_replica2/ has no replicas
   [junit4]   2> 2598158 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50809/i/v/nodes_used_collection_shard2_replica2/ shard2
   [junit4]   2> 2598159 T4802 oascc.SolrZkClient.makePath makePath: /collections/nodes_used_collection/leaders/shard2
   [junit4]   2> 2598172 T4993 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node18",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica3",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50806_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50806/i/v"}
   [junit4]   2> 2598184 T5057 C842 P50806 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node18core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2598185 T5057 C842 P50806 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica3
   [junit4]   2> 2598187 T4734 oasc.SolrCore.close [nodes_used_collection_shard1_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@4b3a2b0c
   [junit4]   2> 2598189 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 2598189 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 2598205 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2598207 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2598207 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2598208 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2598217 T4734 oasc.SolrCore.closeSearcher [nodes_used_collection_shard1_replica1] Closing main searcher on request.
   [junit4]   2> 2598218 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2598218 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data;done=false>>]
   [junit4]   2> 2598220 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data
   [junit4]   2> 2598220 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data/index;done=false>>]
   [junit4]   2> 2598221 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//nodes_used_collection_shard1_replica1/data/index
   [junit4]   2> 2598221 T4734 oasc.SolrCore.close [awholynewcollection_0_shard5_replica3]  CLOSING SolrCore org.apache.solr.core.SolrCore@3e4f8b4b
   [junit4]   2> 2598266 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=4,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2598267 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2598267 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2598268 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2598268 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 2598273 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard5_replica3] Closing main searcher on request.
   [junit4]   2> 2598274 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2598275 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data/index;done=false>>]
   [junit4]   2> 2598275 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data/index
   [junit4]   2> 2598276 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data;done=false>>]
   [junit4]   2> 2598277 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard5_replica3/data
   [junit4]   2> 2598277 T4734 oasc.SolrCore.close [awholynewcollection_0_shard2_replica3]  CLOSING SolrCore org.apache.solr.core.SolrCore@7c62b1db
   [junit4]   2> 2598287 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2598288 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 2598288 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2598288 T4802 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50809/i/v/awholynewcollection_0_shard2_replica1/
   [junit4]   2> 2598289 T4802 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard2_replica1 url=http://127.0.0.1:50809/i/v START replicas=[http://127.0.0.1:50806/i/v/awholynewcollection_0_shard2_replica3/, http://127.0.0.1:50812/i/v/awholynewcollection_0_shard2_replica5/] nUpdates=100
   [junit4]   2> 2598290 T4802 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard2_replica1 url=http://127.0.0.1:50809/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2598290 T4802 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2598291 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2598291 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50809/i/v/awholynewcollection_0_shard2_replica1/ shard2
   [junit4]   2> 2598291 T4802 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leaders/shard2
   [junit4]   2> ASYNC  NEW_CORE C846 name=awholynewcollection_0_shard2_replica5 org.apache.solr.core.SolrCore@a3d63c8 url=http://127.0.0.1:50812/i/v/awholynewcollection_0_shard2_replica5 node=127.0.0.1:50812_i%2Fv C846_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard2_replica5 props:{state=active, core=awholynewcollection_0_shard2_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2598294 T4855 C846 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard2_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2598305 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2598306 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2598307 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2598307 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2598314 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard2_replica3] Closing main searcher on request.
   [junit4]   2> 2598315 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2598315 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data;done=false>>]
   [junit4]   2> 2598316 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data
   [junit4]   2> 2598316 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data/index;done=false>>]
   [junit4]   2> 2598317 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard2_replica3/data/index
   [junit4]   2> 2598318 T4734 oasc.SolrCore.close [awholynewcollection_0_shard3_replica3]  CLOSING SolrCore org.apache.solr.core.SolrCore@4180afac
   [junit4]   2> 2598328 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2>  C843_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica5 props:{state=recovery_failed, core=awholynewcollection_0_shard4_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2598337 T5058 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2598338 T5058 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2598338 T5058 C843 P50812 oasc.RecoveryStrategy.doRecovery Wait 2.0 seconds before trying to recover again (1)
   [junit4]   2> 2598342 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2598343 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2598344 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2598344 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2598345 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard3_replica3] Closing main searcher on request.
   [junit4]   2> 2598346 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2598346 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data/index;done=false>>]
   [junit4]   2> 2598347 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data/index
   [junit4]   2> 2598348 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data;done=false>>]
   [junit4]   2> 2598348 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard3_replica3/data
   [junit4]   2> 2598349 T4734 oasc.SolrCore.close [awholynewcollection_0_shard1_replica3]  CLOSING SolrCore org.apache.solr.core.SolrCore@6a9d4194
   [junit4]   2> 2598363 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2598364 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2598365 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2598365 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2598366 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard1_replica3] Closing main searcher on request.
   [junit4]   2> 2598367 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2598368 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data;done=false>>]
   [junit4]   2> 2598368 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data
   [junit4]   2> 2598369 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data/index;done=false>>]
   [junit4]   2> 2598370 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard1_replica3/data/index
   [junit4]   2> 2598371 T4734 oasc.SolrCore.close [awholynewcollection_0_shard4_replica3]  CLOSING SolrCore org.apache.solr.core.SolrCore@adeb916
   [junit4]   2> 2598390 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2598391 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2598392 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2598393 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2598393 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard4_replica3] Closing main searcher on request.
   [junit4]   2> 2598394 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2598395 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data/index;done=false>>]
   [junit4]   2> 2598395 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data/index
   [junit4]   2> 2598396 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data;done=false>>]
   [junit4]   2> 2598397 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty2-1373800444194//awholynewcollection_0_shard4_replica3/data
   [junit4]   2> 2598398 T4993 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90033385638330376-127.0.0.1:50806_i%2Fv-n_0000000002) am no longer a leader.
   [junit4]   2> 2598403 T4816 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 2598404 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 2598411 T4802 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 2598418 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2598422 T4802 oasc.Overseer.start Overseer (id=90033385638330378-127.0.0.1:50809_i%2Fv-n_0000000003) starting
   [junit4]   2> 2598428 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2598430 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2598430 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2598433 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2598434 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2598439 T4734 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/i/v,null}
   [junit4]   2> 2598443 T4802 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 2598442 T5065 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 2598446 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 2598448 T5064 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 2598457 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard5
   [junit4]   2> 2598461 T5064 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2598465 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2598465 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 2598465 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 2598466 T4802 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50809/i/v/awholynewcollection_0_shard5_replica1/
   [junit4]   2> 2598466 T4802 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard5_replica1 url=http://127.0.0.1:50809/i/v START replicas=[http://127.0.0.1:50812/i/v/awholynewcollection_0_shard5_replica5/] nUpdates=100
   [junit4]   2> 2598468 T4802 oasu.PeerSync.sync PeerSync: core=awholynewcollection_0_shard5_replica1 url=http://127.0.0.1:50809/i/v DONE.  We have no versions.  sync failed.
   [junit4]   2> 2598468 T4802 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 2598468 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 2598469 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50809/i/v/awholynewcollection_0_shard5_replica1/ shard5
   [junit4]   2> 2598469 T4802 oascc.SolrZkClient.makePath makePath: /collections/awholynewcollection_0/leaders/shard5
   [junit4]   2> ASYNC  NEW_CORE C847 name=awholynewcollection_0_shard5_replica5 org.apache.solr.core.SolrCore@47d776f1 url=http://127.0.0.1:50812/i/v/awholynewcollection_0_shard5_replica5 node=127.0.0.1:50812_i%2Fv C847_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard5_replica5 props:{state=active, core=awholynewcollection_0_shard5_replica5, node_name=127.0.0.1:50812_i%2Fv, base_url=http://127.0.0.1:50812/i/v}
   [junit4]   2> 2598471 T4859 C847 P50812 oasc.SolrCore.execute [awholynewcollection_0_shard5_replica5] webapp=/i/v path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 2598492 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2598500 T4734 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50809
   [junit4]   2> 2598500 T4734 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=850965603
   [junit4]   2> 2598503 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2598504 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2598504 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2598507 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2598508 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2598559 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2598560 T5064 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 2598568 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2598569 T5064 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2> 2598577 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node11",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2598578 T5064 oasc.Overseer$ClusterStateUpdater.updateState shard=shard3 is already registered
   [junit4]   2> 2598586 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2598587 T5064 oasc.Overseer$ClusterStateUpdater.updateState shard=shard4 is already registered
   [junit4]   2> 2598595 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node21",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2598595 T5064 oasc.Overseer$ClusterStateUpdater.updateState shard=shard5 is already registered
   [junit4]   2> 2598604 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica2",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2598610 T5064 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2>  C841_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica1 props:{state=recovery_failed, core=awholynewcollection_0_shard4_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2599160 T5061 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2599160 T5061 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2599161 T5061 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2599161 T5061 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2599161 T5061 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2599164 T5061 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2599165 T5061 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2599169 T5067 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2599173 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2599180 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2599180 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2599181 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2599183 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2599184 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2599339 T5058 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2599346 T5068 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2599353 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2599363 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2599364 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2599365 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2599367 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2599368 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2599703 T4788 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2599703 T4788 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 2599704 T4788 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2600176 T5064 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2600179 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2>  C841_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica1 props:{state=down, core=awholynewcollection_0_shard4_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2600208 T5067 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2600209 T5067 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2600209 T5067 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2600209 T5067 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2600210 T5067 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2600214 T5067 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2600216 T5067 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2600223 T5069 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2600228 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2600233 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2600234 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2600234 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2600236 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2600236 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2600373 T5068 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2600373 T5068 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2600374 T5068 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2600374 T5068 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2600375 T5068 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2600377 T5068 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2600377 T5068 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2600378 T5070 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2600384 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2600389 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2600389 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2600390 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2600391 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2600392 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2601253 T5069 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2601254 T5069 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2601254 T5069 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2601254 T5069 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2601255 T5069 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2601257 T5069 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2601257 T5069 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2601258 T5071 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2601262 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2601266 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2601267 T4802 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2601267 T4802 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2601270 T4802 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2601270 T4802 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2601406 T5070 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2601407 T5070 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2601407 T5070 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2601408 T5070 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2601408 T5070 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2601410 T5070 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2601411 T5070 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2601412 T5072 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2601416 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2601427 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2601427 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2601428 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2601430 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2601430 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2601536 T4734 oasc.ZkController.publish publishing core=nodes_used_collection_shard2_replica2 state=down
   [junit4]   2> 2601539 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica1 state=down
   [junit4]   2> 2601541 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica1 state=down
   [junit4]   2> 2601543 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica1 state=down
   [junit4]   2> 2601545 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=down
   [junit4]   2> 2601548 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica1 state=down
   [junit4]   2> 2601733 T5064 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2601736 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2601750 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2601776 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2601794 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2601809 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"nodes_used_collection_shard2_replica2",
   [junit4]   2> 	  "collection":"nodes_used_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2601818 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2601832 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2601844 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node21",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard5",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard5_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2601853 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2601862 T5064 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node11",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard3_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2>  C841_STATE=coll:awholynewcollection_0 core:awholynewcollection_0_shard4_replica1 props:{state=recovery_failed, core=awholynewcollection_0_shard4_replica1, node_name=127.0.0.1:50809_i%2Fv, base_url=http://127.0.0.1:50809/i/v}
   [junit4]   2> 2602286 T5071 C841 P50809 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica1:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2602287 T5071 C841 P50809 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2602287 T5071 C841 P50809 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2602288 T5071 C841 P50809 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2602288 T5071 C841 P50809 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica1 state=recovery_failed
   [junit4]   2> 2602290 T5071 C841 P50809 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2602291 T5071 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2602292 T4734 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node16core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2602292 T5073 C841 P50809 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica1 recoveringAfterStartup=false
   [junit4]   2> 2602293 T5073 C841 P50809 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica1
   [junit4]   2> 2602294 T4734 oasc.SolrCore.close [nodes_used_collection_shard2_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@14e7053b
   [junit4]   2> 2602296 T4802 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2602304 T4802 oasc.LeaderElector$1.process WARN  org.apache.solr.common.SolrException: Fatal Error, SolrCore not found:awholynewcollection_0_shard4_replica1 in []
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:163)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.access$000(LeaderElector.java:55)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector$1.process(LeaderElector.java:129)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:519)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:495)
   [junit4]   2> 
   [junit4]   2> 2602305 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 2602312 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2602313 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2602314 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2602314 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2602315 T4734 oasc.SolrCore.closeSearcher [nodes_used_collection_shard2_replica2] Closing main searcher on request.
   [junit4]   2> 2602316 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2602317 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data;done=false>>]
   [junit4]   2> 2602317 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data
   [junit4]   2> 2602318 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data/index;done=false>>]
   [junit4]   2> 2602319 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//nodes_used_collection_shard2_replica2/data/index
   [junit4]   2> 2602319 T4734 oasc.SolrCore.close [awholynewcollection_0_shard1_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@26d27b42
   [junit4]   2> 2602333 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2602334 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2602335 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2602336 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2602337 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard1_replica1] Closing main searcher on request.
   [junit4]   2> 2602338 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2602339 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data/index;done=false>>]
   [junit4]   2> 2602339 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data/index
   [junit4]   2> 2602340 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data;done=false>>]
   [junit4]   2> 2602341 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard1_replica1/data
   [junit4]   2> 2602341 T4734 oasc.SolrCore.close [awholynewcollection_0_shard2_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@7681485
   [junit4]   2> 2602360 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2602361 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2602362 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2602365 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2602366 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard2_replica1] Closing main searcher on request.
   [junit4]   2> 2602367 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2602368 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data;done=false>>]
   [junit4]   2> 2602368 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data
   [junit4]   2> 2602369 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data/index;done=false>>]
   [junit4]   2> 2602370 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard2_replica1/data/index
   [junit4]   2> 2602370 T4734 oasc.SolrCore.close [awholynewcollection_0_shard5_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@35e3126
   [junit4]   2> 2602384 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2602385 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2602386 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2602386 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2602387 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard5_replica1] Closing main searcher on request.
   [junit4]   2> 2602388 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2602388 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data/index;done=false>>]
   [junit4]   2> 2602389 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data/index
   [junit4]   2> 2602390 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data;done=false>>]
   [junit4]   2> 2602390 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard5_replica1/data
   [junit4]   2> 2602391 T4734 oasc.SolrCore.close [awholynewcollection_0_shard4_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@50d23b4b
   [junit4]   2> 2602404 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2602405 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2602405 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2602406 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2602407 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard4_replica1] Closing main searcher on request.
   [junit4]   2> 2602407 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2602408 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data/index;done=false>>]
   [junit4]   2> 2602408 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data/index
   [junit4]   2> 2602409 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data;done=false>>]
   [junit4]   2> 2602410 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard4_replica1/data
   [junit4]   2> 2602410 T4734 oasc.SolrCore.close [awholynewcollection_0_shard3_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@27a200a0
   [junit4]   2> 2602431 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2602432 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2602433 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2602433 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2602434 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard3_replica1] Closing main searcher on request.
   [junit4]   2> 2602435 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2602435 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data;done=false>>]
   [junit4]   2> 2602436 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data
   [junit4]   2> 2602437 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data/index;done=false>>]
   [junit4]   2> 2602437 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty3-1373800446558//awholynewcollection_0_shard3_replica1/data/index
   [junit4]   2> 2602439 T5064 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90033385638330378-127.0.0.1:50809_i%2Fv-n_0000000003) am no longer a leader.
   [junit4]   2> 2602449 T5072 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2602450 T5072 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2602450 T5072 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2602450 T5072 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2602451 T5072 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2602455 T4734 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/i/v,null}
   [junit4]   2> 2602457 T5072 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2602458 T5072 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2602459 T5074 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2602464 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2602470 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2602471 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2602471 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2602473 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2602473 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2602508 T4734 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50812
   [junit4]   2> 2602509 T4734 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1714793150
   [junit4]   2> 2603497 T5074 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2603498 T5074 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2603498 T5074 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2603498 T5074 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2603498 T5074 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2603501 T5074 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2603501 T5074 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2603502 T5075 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2603506 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2603511 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2603512 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2603512 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2603514 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2603514 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2603742 T4802 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2603743 T4802 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 2603744 T4802 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2604529 T5075 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2604529 T5075 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2604530 T5075 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2604530 T5075 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2604530 T5075 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2604532 T5075 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2604533 T5075 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2604534 T5076 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2604538 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2604543 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2604543 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2604544 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2604550 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2604551 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2605568 T5076 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2605569 T5076 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2605569 T5076 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2605569 T5076 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2605570 T5076 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2605575 T5076 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2605575 T5076 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2605576 T5077 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2605580 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2605584 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2605585 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2605585 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2605587 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2605587 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2606617 T5077 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2606617 T5077 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2606617 T5077 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2606618 T5077 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2606618 T5077 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2606623 T5077 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2606623 T5077 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2606625 T5078 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2606629 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2606634 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2606634 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2606635 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2606637 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2606637 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2607651 T5078 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2607652 T5078 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2607652 T5078 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2607653 T5078 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2607653 T5078 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2607656 T5078 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2607657 T5078 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2607658 T5079 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2607662 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2607667 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2607667 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2607667 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2607669 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2607670 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2608688 T5079 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2608689 T5079 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2608691 T5079 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2608691 T5079 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2608691 T5079 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2608694 T5079 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2608695 T5079 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2608696 T5080 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2608702 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2608708 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2608709 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2608710 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2608712 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2608713 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2609724 T5080 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2609725 T5080 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2609726 T5080 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2609726 T5080 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2609727 T5080 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2609730 T5080 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2609731 T5080 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2609732 T5081 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2609736 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2609741 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2609741 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2609742 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2609744 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2609744 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2610763 T5081 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2610764 T5081 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2610765 T5081 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2610765 T5081 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2610766 T5081 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2610769 T5081 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2610769 T5081 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2610771 T5082 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2610777 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2610784 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2610785 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2610786 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2610789 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2610789 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2611807 T5082 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2611807 T5082 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2611808 T5082 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2611808 T5082 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2611808 T5082 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2611811 T5082 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2611811 T5082 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2611813 T5083 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2611820 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2611826 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2611827 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2611827 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2611829 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2611830 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2612855 T5083 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2612856 T5083 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2612857 T5083 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2612857 T5083 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2612858 T5083 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2612861 T5083 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2612863 T5083 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2612873 T5084 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2612878 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2612882 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2612883 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2612883 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2612885 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2612886 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2613916 T5084 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2613917 T5084 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2613917 T5084 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2613917 T5084 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2613918 T5084 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2613920 T5084 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2613921 T5084 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2613922 T5086 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2613926 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2613932 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2613932 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2613933 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2613935 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2613936 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2614950 T5086 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2614951 T5086 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2614952 T5086 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2614952 T5086 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2614952 T5086 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2614979 T5086 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2614980 T5086 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2614982 T5087 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2615009 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2615031 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2615032 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2615033 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2615036 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2615037 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2616040 T5087 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2616041 T5087 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2616041 T5087 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2616042 T5087 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2616042 T5087 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2616044 T5087 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2616045 T5087 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2616046 T5088 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2616050 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2616055 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2616055 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2616056 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2616058 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2616059 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2617079 T5088 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2617081 T5088 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2617082 T5088 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2617084 T5088 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2617085 T5088 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2617088 T5088 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2617088 T5088 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2617089 T5089 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2617093 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2617101 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2617102 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2617102 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2617105 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2617105 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2618119 T5089 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2618120 T5089 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2618120 T5089 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2618120 T5089 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2618121 T5089 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2618123 T5089 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2618124 T5089 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2618125 T5090 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2618130 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2618135 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2618136 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2618137 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2618139 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2618140 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2619175 T5090 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2619175 T5090 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2619176 T5090 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2619176 T5090 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2619176 T5090 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2619179 T5090 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2619179 T5090 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2619181 T5091 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2619185 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2619192 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2619193 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2619194 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2619196 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2619196 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2620217 T5091 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2620217 T5091 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2620217 T5091 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2620218 T5091 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2620218 T5091 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2620221 T5091 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2620222 T5091 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2620223 T5092 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2620231 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2620252 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2620253 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2620254 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2620256 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2620260 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2621250 T5092 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2621250 T5092 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2621251 T5092 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2621251 T5092 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2621252 T5092 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2621255 T5092 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2621256 T5092 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2621257 T5093 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2621262 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2621269 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2621270 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2621271 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2621275 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2621276 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2622291 T5093 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2622292 T5093 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2622292 T5093 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2622292 T5093 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2622293 T5093 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2622295 T5093 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2622296 T5093 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2622297 T5094 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2622307 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2622317 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2622327 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2622330 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2622335 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2622339 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2623337 T5094 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2623338 T5094 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2623338 T5094 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2623339 T5094 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2623339 T5094 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2623341 T5094 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2623342 T5094 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2623343 T5095 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2623348 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2623353 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2623354 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2623355 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2623357 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2623358 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2624381 T5095 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2624382 T5095 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2624382 T5095 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2624382 T5095 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2624382 T5095 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2624385 T5095 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2624385 T5095 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2624387 T5096 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2624391 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2624396 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2624397 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2624397 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2624399 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2624400 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2625433 T5096 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2625434 T5096 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2625434 T5096 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2625434 T5096 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2625435 T5096 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2625437 T5096 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2625438 T5096 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2625439 T5097 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2625443 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2625449 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2625450 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2625451 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2625458 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2625459 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2626469 T5097 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2626470 T5097 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2626471 T5097 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2626471 T5097 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2626472 T5097 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2626483 T5097 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2626483 T5097 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2626485 T5098 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2626492 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2626499 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2626500 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2626501 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2626504 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2626506 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2627485 T5098 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2627486 T5098 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2627487 T5098 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2627487 T5098 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2627487 T5098 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2627489 T5098 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2627490 T5098 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2627491 T5099 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2627495 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2627500 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2627501 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2627501 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2627504 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2627505 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2628530 T5099 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2628531 T5099 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2628531 T5099 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2628531 T5099 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2628532 T5099 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2628534 T5099 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2628535 T5099 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2628536 T5100 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2628540 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2628546 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2628547 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2628547 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2628550 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2628551 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2629576 T5100 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2629577 T5100 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2629577 T5100 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2629577 T5100 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2629577 T5100 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2629579 T5100 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2629580 T5100 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2629581 T5101 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2629587 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2629591 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2629591 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2629592 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2629594 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2629595 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2630588 T5101 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2630589 T5101 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2630589 T5101 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2630589 T5101 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2630590 T5101 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2630593 T5101 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2630593 T5101 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2630595 T5102 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2630599 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2630605 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2630606 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2630606 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2630609 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2630610 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2631642 T5102 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2631643 T5102 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2631644 T5102 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2631644 T5102 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2631645 T5102 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2631648 T5102 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2631648 T5102 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2631649 T5103 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2631657 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2631662 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 2631663 T4816 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was recovery_failed, I won't be the leader.
   [junit4]   2> 2631664 T4816 oasc.ShardLeaderElectionContext.rejoinLeaderElection There may be a better leader candidate than us - going back into recovery
   [junit4]   2> 2631667 T4816 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 2631668 T4816 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2632571 T4734 oasc.ZkController.publishAndWaitForDownStates WARN Timed out waiting to see all nodes published as DOWN in our cluster state.
   [junit4]   2> 2632572 T4734 oasc.ZkController.publish publishing core=nodes_used_collection_shard1_replica2 state=down
   [junit4]   2> 2632575 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard1_replica5 state=down
   [junit4]   2> 2632578 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard3_replica5 state=down
   [junit4]   2> 2632581 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=down
   [junit4]   2> 2632584 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard2_replica5 state=down
   [junit4]   2> 2632587 T4734 oasc.ZkController.publish publishing core=awholynewcollection_0_shard5_replica5 state=down
   [junit4]   2> 2632688 T5103 C843 P50812 oasc.SolrException.log ERROR Error while trying to recover. core=awholynewcollection_0_shard4_replica5:org.apache.solr.common.SolrException: No registered leader was found, collection:awholynewcollection_0 slice:shard4
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:487)
   [junit4]   2> 		at org.apache.solr.common.cloud.ZkStateReader.getLeaderRetry(ZkStateReader.java:470)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.doRecovery(RecoveryStrategy.java:317)
   [junit4]   2> 		at org.apache.solr.cloud.RecoveryStrategy.run(RecoveryStrategy.java:219)
   [junit4]   2> 	
   [junit4]   2> 2632688 T5103 C843 P50812 oasc.RecoveryStrategy.doRecovery ERROR Recovery failed - trying again... (0) core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2632689 T5103 C843 P50812 oasc.SolrException.log ERROR Recovery failed - interrupted. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2632689 T5103 C843 P50812 oasc.SolrException.log ERROR Recovery failed - I give up. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2632689 T5103 C843 P50812 oasc.ZkController.publish publishing core=awholynewcollection_0_shard4_replica5 state=recovery_failed
   [junit4]   2> 2632692 T5103 C843 P50812 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2632692 T5103 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2632693 T4734 oasc.RecoveryStrategy.close WARN Stopping recovery for zkNodeName=core_node20core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2632694 T5104 C843 P50812 oasc.RecoveryStrategy.run Starting recovery process.  core=awholynewcollection_0_shard4_replica5 recoveringAfterStartup=false
   [junit4]   2> 2632695 T5104 C843 P50812 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=awholynewcollection_0_shard4_replica5
   [junit4]   2> 2632695 T4734 oasc.SolrCore.close [nodes_used_collection_shard1_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@2c000624
   [junit4]   2> 2632698 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 2632704 T4816 oasc.LeaderElector$1.process WARN  org.apache.solr.common.SolrException: Fatal Error, SolrCore not found:awholynewcollection_0_shard4_replica5 in []
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:163)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.joinElection(LeaderElector.java:266)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.rejoinLeaderElection(ElectionContext.java:378)
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:170)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.access$000(LeaderElector.java:55)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector$1.process(LeaderElector.java:129)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:519)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:495)
   [junit4]   2> 
   [junit4]   2> 2632711 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 2632713 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=1,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2632714 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2632715 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2632715 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2632720 T4734 oasc.SolrCore.closeSearcher [nodes_used_collection_shard1_replica2] Closing main searcher on request.
   [junit4]   2> 2632722 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2632723 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data;done=false>>]
   [junit4]   2> 2632723 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data
   [junit4]   2> 2632724 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data/index;done=false>>]
   [junit4]   2> 2632725 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//nodes_used_collection_shard1_replica2/data/index
   [junit4]   2> 2632725 T4734 oasc.SolrCore.close [awholynewcollection_0_shard1_replica5]  CLOSING SolrCore org.apache.solr.core.SolrCore@4f6f9093
   [junit4]   2> 2632736 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard5
   [junit4]   2> 2632749 T4816 oasc.LeaderElector$1.process WARN  org.apache.solr.common.SolrException: Fatal Error, SolrCore not found:awholynewcollection_0_shard5_replica5 in []
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:163)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.access$000(LeaderElector.java:55)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector$1.process(LeaderElector.java:129)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:519)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:495)
   [junit4]   2> 
   [junit4]   2> 2632751 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=4,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2632752 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2632753 T4816 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 2632753 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2632754 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2632757 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard1_replica5] Closing main searcher on request.
   [junit4]   2> 2632758 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2632758 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data;done=false>>]
   [junit4]   2> 2632759 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data
   [junit4]   2> 2632760 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data/index;done=false>>]
   [junit4]   2> 2632760 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard1_replica5/data/index
   [junit4]   2> 2632759 T4816 oasc.LeaderElector$1.process WARN  org.apache.solr.common.SolrException: Fatal Error, SolrCore not found:awholynewcollection_0_shard2_replica5 in []
   [junit4]   2> 	at org.apache.solr.cloud.ShardLeaderElectionContext.runLeaderProcess(ElectionContext.java:163)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.runIamLeaderProcess(LeaderElector.java:156)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.checkIfIamLeader(LeaderElector.java:100)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector.access$000(LeaderElector.java:55)
   [junit4]   2> 	at org.apache.solr.cloud.LeaderElector$1.process(LeaderElector.java:129)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:519)
   [junit4]   2> 	at org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:495)
   [junit4]   2> 
   [junit4]   2> 2632761 T4734 oasc.SolrCore.close [awholynewcollection_0_shard3_replica5]  CLOSING SolrCore org.apache.solr.core.SolrCore@5aad7202
   [junit4]   2> 2632764 T4816 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 2632769 T4816 oasc.Overseer.start Overseer (id=90033385638330380-127.0.0.1:50812_i%2Fv-n_0000000004) starting
   [junit4]   2> 2632779 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2632780 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2632781 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2632781 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2632783 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard3_replica5] Closing main searcher on request.
   [junit4]   2> 2632784 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2632784 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data/index;done=false>>]
   [junit4]   2> 2632785 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data/index
   [junit4]   2> 2632791 T5106 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 2632794 T4816 oascc.ZkStateReader$3.process Updating live nodes... (1)
   [junit4]   2> 2632794 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 2632798 T5105 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 2632800 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data;done=false>>]
   [junit4]   2> 2632801 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard3_replica5/data
   [junit4]   2> 2632808 T5105 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 2632811 T4734 oasc.SolrCore.close [awholynewcollection_0_shard4_replica5]  CLOSING SolrCore org.apache.solr.core.SolrCore@57b95f5d
   [junit4]   2> 2632813 T5105 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node16",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica1",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50809_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50809/i/v"}
   [junit4]   2> 2632825 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2632826 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2632826 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2632827 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2632828 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard4_replica5] Closing main searcher on request.
   [junit4]   2> 2632829 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2632829 T5105 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=5 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node20",
   [junit4]   2> 	  "numShards":"5",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovery_failed",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard4_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2632830 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data/index;done=false>>]
   [junit4]   2> 2632830 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data/index
   [junit4]   2> 2632831 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data;done=false>>]
   [junit4]   2> 2632832 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard4_replica5/data
   [junit4]   2> 2632832 T4734 oasc.SolrCore.close [awholynewcollection_0_shard2_replica5]  CLOSING SolrCore org.apache.solr.core.SolrCore@a3d63c8
   [junit4]   2> 2632846 T5105 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard1_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2632846 T5105 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 2632854 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2632855 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2632855 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2632856 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2632857 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard2_replica5] Closing main searcher on request.
   [junit4]   2> 2632858 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2632859 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data;done=false>>]
   [junit4]   2> 2632859 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data
   [junit4]   2> 2632860 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data/index;done=false>>]
   [junit4]   2> 2632860 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard2_replica5/data/index
   [junit4]   2> 2632861 T4734 oasc.SolrCore.close [awholynewcollection_0_shard5_replica5]  CLOSING SolrCore org.apache.solr.core.SolrCore@47d776f1
   [junit4]   2> 2632884 T4734 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 2632885 T4734 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 2632886 T4734 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 2632886 T4734 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 2632887 T5105 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node9",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"awholynewcollection_0_shard2_replica5",
   [junit4]   2> 	  "collection":"awholynewcollection_0",
   [junit4]   2> 	  "node_name":"127.0.0.1:50812_i%2Fv",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50812/i/v"}
   [junit4]   2> 2632888 T5105 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2> 2632888 T4734 oasc.SolrCore.closeSearcher [awholynewcollection_0_shard5_replica5] Closing main searcher on request.
   [junit4]   2> 2632890 T4734 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 2632890 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data;done=false>>]
   [junit4]   2> 2632891 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data
   [junit4]   2> 2632894 T4734 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data/index;done=false>>]
   [junit4]   2> 2632894 T4734 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.CollectionsAPIDistributedZkTest-jetty4-1373800448749//awholynewcollection_0_shard5_replica5/data/index
   [junit4]   2> 2632898 T4816 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2632898 T4816 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 2632899 T4816 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 2632912 T4734 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/i/v,null}
   [junit4]   2> 2632971 T4734 oas.SolrTestCaseJ4.tearDown ###Ending testDistribSearch
   [junit4]   2> 2632975 T4734 oasc.ZkTestServer.send4LetterWord connecting to 127.0.0.1:50796 50796
   [junit4]   1> client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=CollectionsAPIDistributedZkTest -Dtests.method=testDistribSearch -Dtests.seed=DFF894EA9CA91D54 -Dtests.slow=true -Dtests.locale=sv_SE -Dtests.timezone=Australia/NSW -Dtests.file.encoding=US-ASCII
   [junit4] ERROR    127s | CollectionsAPIDistributedZkTest.testDistribSearch <<<
   [junit4]    > Throwable #1: org.apache.solr.client.solrj.impl.HttpSolrServer$RemoteSolrException: collection already exists: awholynewcollection_0
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([DFF894EA9CA91D54:5E1E1AF2EBF67D68]:0)
   [junit4]    > 	at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:424)
   [junit4]    > 	at org.apache.solr.client.solrj.impl.HttpSolrServer.request(HttpSolrServer.java:180)
   [junit4]    > 	at org.apache.solr.client.solrj.impl.LBHttpSolrServer.request(LBHttpSolrServer.java:264)
   [junit4]    > 	at org.apache.solr.client.solrj.impl.CloudSolrServer.request(CloudSolrServer.java:318)
   [junit4]    > 	at org.apache.solr.cloud.AbstractFullDistribZkTestBase.createCollection(AbstractFullDistribZkTestBase.java:1522)
   [junit4]    > 	at org.apache.solr.cloud.CollectionsAPIDistributedZkTest.testCollectionsAPI(CollectionsAPIDistributedZkTest.java:438)
   [junit4]    > 	at org.apache.solr.cloud.CollectionsAPIDistributedZkTest.doTest(CollectionsAPIDistributedZkTest.java:146)
   [junit4]    > 	at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:835)
   [junit4]    > 	at java.lang.Thread.run(Thread.java:724)
   [junit4]   2> 2634145 T4734 oas.SolrTestCaseJ4.deleteCore ###deleteCore
   [junit4]   2> NOTE: test params are: codec=Lucene42: {}, docValues:{}, sim=RandomSimilarityProvider(queryNorm=true,coord=crazy): {}, locale=sv_SE, timezone=Australia/NSW
   [junit4]   2> NOTE: Mac OS X 10.8.4 x86_64/Oracle Corporation 1.7.0_25 (64-bit)/cpus=2,threads=2,free=231274776,total=536870912
   [junit4]   2> NOTE: All tests run in this JVM: [TestSchemaSimilarityResource, CurrencyFieldOpenExchangeTest, TestPropInjectDefaults, TestWordDelimiterFilterFactory, QueryElevationComponentTest, CoreAdminHandlerTest, TestFieldCollectionResource, OverseerTest, BlockDirectoryTest, QueryResultKeyTest, SignatureUpdateProcessorFactoryTest, URLClassifyProcessorTest, TestStressRecovery, ReturnFieldsTest, ConvertedLegacyTest, TestSchemaVersionResource, TestUpdate, TestQuerySenderListener, SpellPossibilityIteratorTest, TestLMDirichletSimilarityFactory, TestFaceting, TestCSVResponseWriter, TestPHPSerializedResponseWriter, TestDistributedGrouping, TestMultiCoreConfBootstrap, TestSolrDeletionPolicy1, RequiredFieldsTest, DirectUpdateHandlerTest, TestConfig, TestFieldResource, ZkControllerTest, AutoCommitTest, SolrCoreCheckLockOnStartupTest, TestArbitraryIndexDir, TestGroupingSearch, ZkSolrClientTest, HdfsUnloadDistributedZkTest, DateMathParserTest, SyncSliceTest, CacheHeaderTest, HdfsLockFactoryTest, TestDynamicFieldResource, RecoveryZkTest, ResourceLoaderTest, SolrTestCaseJ4Test, PrimUtilsTest, ResponseLogComponentTest, TestSolrQueryParser, SuggesterWFSTTest, HdfsBasicDistributedZk2Test, TestCloudManagedSchema, DefaultValueUpdateProcessorTest, TestCloudManagedSchemaAddField, TestIndexSearcher, TestStressReorder, HdfsChaosMonkeySafeLeaderTest, RAMDirectoryFactoryTest, BlockCacheTest, TestLuceneMatchVersion, ShardSplitTest, TestRangeQuery, TestDefaultSimilarityFactory, TestRecovery, HighlighterTest, CircularListTest, DocValuesTest, TestDynamicFieldCollectionResource, SpatialFilterTest, TestZkChroot, DocumentBuilderTest, ParsingFieldUpdateProcessorsTest, StatelessScriptUpdateProcessorFactoryTest, SimplePostToolTest, AliasIntegrationTest, TestFuzzyAnalyzedSuggestions, SimpleFacetsTest, TestReload, TestXIncludeConfig, TestSolr4Spatial, SliceStateTest, LegacyHTMLStripCharFilterTest, WordBreakSolrSpellCheckerTest, TestSolrXmlPersistence, TestUniqueKeyFieldResource, TestSerializedLuceneMatchVersion, IndexSchemaTest, PreAnalyzedUpdateProcessorTest, EchoParamsTest, TestClassNameShortening, LeaderElectionIntegrationTest, TestSolrDeletionPolicy2, TestCSVLoader, DistributedSpellCheckComponentTest, SuggesterTSTTest, TestRecoveryHdfs, CoreAdminCreateDiscoverTest, TestFastOutputStream, TestJmxMonitoredMap, BasicHdfsTest, RequestHandlersTest, TestLRUCache, SpellingQueryConverterTest, TestQueryUtils, TestSearchPerf, TestCopyFieldCollectionResource, DocumentAnalysisRequestHandlerTest, TestDocSet, TestLMJelinekMercerSimilarityFactory, NoCacheHeaderTest, TestSurroundQueryParser, ShowFileRequestHandlerTest, HdfsRecoveryZkTest, ClusterStateUpdateTest, TestSchemaNameResource, TestSolrIndexConfig, LeaderElectionTest, ZkNodePropsTest, TestManagedSchema, HdfsDirectoryTest, PathHierarchyTokenizerFactoryTest, TestCoreDiscovery, BasicFunctionalityTest, HdfsBasicDistributedZkTest, TestQuerySenderNoQuery, XsltUpdateRequestHandlerTest, CurrencyFieldXmlFileTest, TestFastLRUCache, TestValueSourceCache, HdfsCollectionsAPIDistributedZkTest, NumericFieldsTest, AddSchemaFieldsUpdateProcessorFactoryTest, DistributedTermsComponentTest, TestTrie, TestRandomDVFaceting, MBeansHandlerTest, TestDistributedSearch, SolrIndexConfigTest, TestSweetSpotSimilarityFactory, TestMergePolicyConfig, TestCodecSupport, TestIndexingPerformance, TestPseudoReturnFields, TestMaxScoreQueryParser, QueryParsingTest, DirectSolrConnectionTest, DirectUpdateHandlerOptimizeTest, TestFiltering, BasicDistributedZkTest, ChaosMonkeySafeLeaderTest, BasicDistributedZk2Test, UnloadDistributedZkTest, CollectionsAPIDistributedZkTest]
   [junit4] Completed in 126.77s, 1 test, 1 error <<< FAILURES!

[...truncated 483 lines...]
BUILD FAILED
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:389: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:369: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:39: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build.xml:181: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/common-build.xml:449: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/lucene/common-build.xml:1247: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/lucene/common-build.xml:890: There were test failures: 318 suites, 1342 tests, 1 error, 39 ignored (5 assumptions)

Total time: 108 minutes 59 seconds
Build step 'Invoke Ant' marked build as failure
Description set: Java: 64bit/jdk1.7.0 -XX:+UseCompressedOops -XX:+UseG1GC
Archiving artifacts
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure



[JENKINS] Lucene-Solr-trunk-MacOSX (64bit/jdk1.7.0) - Build # 648 - Still Failing!

Posted by Policeman Jenkins Server <je...@thetaphi.de>.
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-MacOSX/648/
Java: 64bit/jdk1.7.0 -XX:+UseCompressedOops -XX:+UseParallelGC

2 tests failed.
REGRESSION:  org.apache.solr.cloud.ShardSplitTest.testDistribSearch

Error Message:
shard1_1 is not consistent expected:<81> but was:<80>

Stack Trace:
java.lang.AssertionError: shard1_1 is not consistent expected:<81> but was:<80>
	at __randomizedtesting.SeedInfo.seed([30BF22C5938FEB5:82ED7C342E679E89]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.junit.Assert.failNotEquals(Assert.java:647)
	at org.junit.Assert.assertEquals(Assert.java:128)
	at org.junit.Assert.assertEquals(Assert.java:472)
	at org.apache.solr.cloud.ShardSplitTest.checkSubShardConsistency(ShardSplitTest.java:247)
	at org.apache.solr.cloud.ShardSplitTest.checkDocCountsAndShardStates(ShardSplitTest.java:207)
	at org.apache.solr.cloud.ShardSplitTest.doTest(ShardSplitTest.java:173)
	at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:835)
	at sun.reflect.GeneratedMethodAccessor64.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
	at org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
	at java.lang.Thread.run(Thread.java:724)


REGRESSION:  org.apache.solr.schema.TestCloudManagedSchemaAddField.testDistribSearch

Error Message:
Max retry count 20 exceeded after 734 ms.  QUERY FAILED: xpath=/response/lst[@name='responseHeader']/int[@name='status'][.='0']  request=/schema/fields/newfield5?wt=xml  response=<?xml version="1.0" encoding="UTF-8"?> <response>  <lst name="responseHeader">   <int name="status">404</int>   <int name="QTime">1</int> </lst> <lst name="error">   <str name="msg">Field 'newfield5' not found.</str>   <int name="code">404</int> </lst> </response> 

Stack Trace:
java.lang.AssertionError: Max retry count 20 exceeded after 734 ms.  QUERY FAILED: xpath=/response/lst[@name='responseHeader']/int[@name='status'][.='0']  request=/schema/fields/newfield5?wt=xml  response=<?xml version="1.0" encoding="UTF-8"?>
<response>

<lst name="responseHeader">
  <int name="status">404</int>
  <int name="QTime">1</int>
</lst>
<lst name="error">
  <str name="msg">Field 'newfield5' not found.</str>
  <int name="code">404</int>
</lst>
</response>

	at __randomizedtesting.SeedInfo.seed([30BF22C5938FEB5:82ED7C342E679E89]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.apache.solr.schema.TestCloudManagedSchemaAddField.doTest(TestCloudManagedSchemaAddField.java:119)
	at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:835)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1559)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.access$600(RandomizedRunner.java:79)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:737)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:773)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:787)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
	at org.apache.lucene.util.TestRuleFieldCacheSanity$1.evaluate(TestRuleFieldCacheSanity.java:51)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:782)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:442)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:746)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:648)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:682)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:693)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:53)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesInvariantRule$1.evaluate(SystemPropertiesInvariantRule.java:55)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:43)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:70)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:358)
	at java.lang.Thread.run(Thread.java:724)




Build Log:
[...truncated 9545 lines...]
   [junit4] Suite: org.apache.solr.schema.TestCloudManagedSchemaAddField
   [junit4]   2> 1771781 T3930 oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system property: /
   [junit4]   2> 1771795 T3930 oas.SolrTestCaseJ4.setUp ###Starting testDistribSearch
   [junit4]   2> Creating dataDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./solrtest-TestCloudManagedSchemaAddField-1373827508992
   [junit4]   2> 1771798 T3930 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   2> 1771799 T3931 oasc.ZkTestServer$ZKServerMain.runFromConfig Starting server
   [junit4]   2> 1771814 T3930 oasc.ZkTestServer.run start zk server on port:50149
   [junit4]   2> 1771835 T3932 oazs.NIOServerCnxn.doIO WARN Exception causing close of session 0x0 due to java.io.IOException: ZooKeeperServer not running
   [junit4]   2> 1771844 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1773201 T3934 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@60608ab5 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1773202 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1773202 T3930 oascc.SolrZkClient.makePath makePath: /solr
   [junit4]   2> 1773224 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1773245 T3939 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@6600fcd5 name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1773246 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1773247 T3930 oascc.SolrZkClient.makePath makePath: /collections/collection1
   [junit4]   2> 1773270 T3930 oascc.SolrZkClient.makePath makePath: /collections/collection1/shards
   [junit4]   2> 1773280 T3930 oascc.SolrZkClient.makePath makePath: /collections/control_collection
   [junit4]   2> 1773286 T3930 oascc.SolrZkClient.makePath makePath: /collections/control_collection/shards
   [junit4]   2> 1773296 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/solrconfig-tlog-mutable-managed-schema.xml to /configs/conf1/solrconfig.xml
   [junit4]   2> 1773298 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/solrconfig.xml
   [junit4]   2> 1773315 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/schema.xml to /configs/conf1/schema.xml
   [junit4]   2> 1773316 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/schema.xml
   [junit4]   2> 1773326 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
   [junit4]   2> 1773327 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/stopwords.txt
   [junit4]   2> 1773338 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
   [junit4]   2> 1773340 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/protwords.txt
   [junit4]   2> 1773349 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
   [junit4]   2> 1773350 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/currency.xml
   [junit4]   2> 1773362 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 1773363 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/open-exchange-rates.json
   [junit4]   2> 1773372 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 1773374 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 1773385 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
   [junit4]   2> 1773387 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/old_synonyms.txt
   [junit4]   2> 1773397 T3930 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
   [junit4]   2> 1773398 T3930 oascc.SolrZkClient.makePath makePath: /configs/conf1/synonyms.txt
   [junit4]   2> 1774083 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1774107 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50153
   [junit4]   2> 1774108 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1774109 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1774110 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604
   [junit4]   2> 1774111 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604/'
   [junit4]   2> 1774158 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604/solr.xml
   [junit4]   2> 1774287 T3930 oasc.CoreContainer.<init> New CoreContainer 1819012486
   [junit4]   2> 1774287 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604/]
   [junit4]   2> 1774289 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1774290 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1774290 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1774291 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1774291 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1774292 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1774292 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1774293 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1774293 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1774294 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1774315 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1774317 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1774319 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1774324 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1774334 T3950 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@47c8e571 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1774335 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1774340 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1774359 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1774364 T3952 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@6b37777e name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1774365 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1774370 T3930 oascc.SolrZkClient.makePath makePath: /overseer/queue
   [junit4]   2> 1774381 T3930 oascc.SolrZkClient.makePath makePath: /overseer/collection-queue-work
   [junit4]   2> 1774393 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes
   [junit4]   2> 1774398 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50153_
   [junit4]   2> 1774405 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50153_
   [junit4]   2> 1774413 T3930 oascc.SolrZkClient.makePath makePath: /overseer_elect/election
   [junit4]   2> 1774429 T3930 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1774435 T3930 oasc.Overseer.start Overseer (id=90035159632510979-127.0.0.1:50153_-n_0000000000) starting
   [junit4]   2> 1774445 T3930 oascc.SolrZkClient.makePath makePath: /overseer/queue-work
   [junit4]   2> 1774456 T3954 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1774459 T3930 oascc.SolrZkClient.makePath makePath: /clusterstate.json
   [junit4]   2> 1774467 T3930 oascc.SolrZkClient.makePath makePath: /aliases.json
   [junit4]   2> 1774477 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1774488 T3953 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1774496 T3955 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1774497 T3955 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1774499 T3955 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 1776002 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1776004 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=1 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"1",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50153_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50153"}
   [junit4]   2> 1776004 T3953 oasc.Overseer$ClusterStateUpdater.createCollection Create collection control_collection with numShards 1
   [junit4]   2> 1776004 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard1
   [junit4]   2> 1776017 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1776501 T3955 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1776502 T3955 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604/collection1
   [junit4]   2> 1776502 T3955 oasc.ZkController.createCollectionZkNode Check for collection zkNode:control_collection
   [junit4]   2> 1776504 T3955 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1776505 T3955 oasc.ZkController.readConfigName Load collection config from:/collections/control_collection
   [junit4]   2> 1776511 T3955 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604/collection1/'
   [junit4]   2> 1776513 T3955 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604/collection1/lib/classes/' to classloader
   [junit4]   2> 1776513 T3955 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604/collection1/lib/README' to classloader
   [junit4]   2> 1776598 T3955 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 1776684 T3955 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1776687 T3955 oass.ManagedIndexSchemaFactory.create The schema is configured as managed, but managed schema resource managed-schema not found - loading non-managed schema schema.xml instead
   [junit4]   2> 1776690 T3955 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 1776707 T3955 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1777573 T3955 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1777585 T3955 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1777589 T3955 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1777628 T3955 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1777636 T3955 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1777656 T3955 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1777660 T3955 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1777660 T3955 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1777661 T3955 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1777663 T3955 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1777664 T3955 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1777674 T3955 oass.ManagedIndexSchema.persistManagedSchemaToZooKeeper Created and persisted managed schema znode at /configs/conf1/managed-schema
   [junit4]   2> 1777678 T3955 oascc.SolrZkClient.makePath makePath: /configs/conf1/schema.xml.bak
   [junit4]   2> 1777692 T3955 oass.ManagedIndexSchemaFactory.zkUgradeToManagedSchema After upgrading to managed schema in ZooKeeper, renamed the non-managed schema /configs/conf1/schema.xml to /configs/conf1/schema.xml.bak
   [junit4]   2> 1777693 T3955 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1777693 T3955 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-controljetty-1373827510604/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data/
   [junit4]   2> 1777693 T3955 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1777694 T3955 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data
   [junit4]   2> 1777695 T3955 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data/index/
   [junit4]   2> 1777695 T3955 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data/index' doesn't exist. Creating new index...
   [junit4]   2> 1777695 T3955 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data/index
   [junit4]   2> 1777700 T3955 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@20672345 lockFactory=org.apache.lucene.store.NativeFSLockFactory@71ee1b26),segFN=segments_1,generation=1}
   [junit4]   2> 1777700 T3955 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1777707 T3955 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1777708 T3955 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1777708 T3955 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1777709 T3955 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1777710 T3955 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1777710 T3955 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1777711 T3955 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1777726 T3955 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1777734 T3955 oass.SolrIndexSearcher.<init> Opening Searcher@4556b6d3 main
   [junit4]   2> 1777735 T3955 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1777736 T3955 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1777736 T3955 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1777749 T3956 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@4556b6d3 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1777750 T3955 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1777751 T3955 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50153 collection:control_collection shard:shard1
   [junit4]   2> 1777751 T3955 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1777774 T3955 oascc.SolrZkClient.makePath makePath: /collections/control_collection/leader_elect/shard1/election
   [junit4]   2> 1777790 T3955 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 1777796 T3955 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 1777796 T3955 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1777796 T3955 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50153/collection1/
   [junit4]   2> 1777797 T3955 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 1777797 T3955 oasc.SyncStrategy.syncToMe http://127.0.0.1:50153/collection1/ has no replicas
   [junit4]   2> 1777797 T3955 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50153/collection1/ shard1
   [junit4]   2> 1777797 T3955 oascc.SolrZkClient.makePath makePath: /collections/control_collection/leaders/shard1
   [junit4]   2> 1779052 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1779074 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1779101 T3955 oasc.ZkController.register We are http://127.0.0.1:50153/collection1/ and leader is http://127.0.0.1:50153/collection1/
   [junit4]   2> 1779102 T3955 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 1779102 T3955 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1779103 T3955 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1779107 T3955 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1779111 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1779111 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1779113 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1779140 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:
   [junit4]   2> 1779144 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1779153 T3959 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@47fe2d18 name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1779154 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1779158 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1779167 T3930 oasc.ChaosMonkey.monkeyLog monkey: init - expire sessions:true cause connection loss:true
   [junit4]   2> 1779758 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1779766 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50157
   [junit4]   2> 1779767 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1779768 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1779768 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363
   [junit4]   2> 1779769 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363/'
   [junit4]   2> 1779819 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363/solr.xml
   [junit4]   2> 1779944 T3930 oasc.CoreContainer.<init> New CoreContainer 1721190892
   [junit4]   2> 1779945 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363/]
   [junit4]   2> 1779947 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1779948 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1779948 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1779949 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1779949 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1779949 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1779950 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1779951 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1779951 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1779952 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1779974 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1779975 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1779976 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1779978 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1779982 T3970 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@4120e752 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1779982 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1779994 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1780014 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1780019 T3972 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@7eeb02ab name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1780020 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1780036 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1780594 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1780597 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=1 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"1",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50153_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50153"}
   [junit4]   2> 1780606 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1780610 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1780613 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1781048 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50157_
   [junit4]   2> 1781052 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50157_
   [junit4]   2> 1781065 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1781068 T3972 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 1781068 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1781068 T3952 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 1781068 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1781072 T3959 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 1781084 T3973 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1781085 T3973 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1781087 T3973 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 1782125 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1782128 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50157_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50157"}
   [junit4]   2> 1782128 T3953 oasc.Overseer$ClusterStateUpdater.createCollection Create collection collection1 with numShards 4
   [junit4]   2> 1782128 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard1
   [junit4]   2> 1782139 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1782142 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1782145 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1783095 T3973 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1783096 T3973 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363/collection1
   [junit4]   2> 1783096 T3973 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 1783098 T3973 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1783098 T3973 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 1783100 T3973 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363/collection1/'
   [junit4]   2> 1783102 T3973 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363/collection1/lib/classes/' to classloader
   [junit4]   2> 1783103 T3973 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363/collection1/lib/README' to classloader
   [junit4]   2> 1783197 T3973 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 1783291 T3973 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1783295 T3973 oass.IndexSchema.readSchema Reading Solr Schema from managed-schema
   [junit4]   2> 1783307 T3973 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1784259 T3973 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1784275 T3973 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1784278 T3973 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1784287 T3973 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1784294 T3973 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1784324 T3973 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1784332 T3973 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1784332 T3973 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1784332 T3973 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1784335 T3973 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1784335 T3973 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1784345 T3973 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1784345 T3973 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty1-1373827516363/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1/
   [junit4]   2> 1784346 T3973 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1784350 T3973 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1
   [junit4]   2> 1784351 T3973 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1/index/
   [junit4]   2> 1784351 T3973 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1/index' doesn't exist. Creating new index...
   [junit4]   2> 1784352 T3973 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1/index
   [junit4]   2> 1784358 T3973 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@64d807a9 lockFactory=org.apache.lucene.store.NativeFSLockFactory@4645c508),segFN=segments_1,generation=1}
   [junit4]   2> 1784358 T3973 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1784370 T3973 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1784372 T3973 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1784373 T3973 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1784373 T3973 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1784374 T3973 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1784375 T3973 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1784375 T3973 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1784394 T3973 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1784413 T3973 oass.SolrIndexSearcher.<init> Opening Searcher@3f8d4e5c main
   [junit4]   2> 1784415 T3973 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1784415 T3973 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1784416 T3973 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1784431 T3974 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@3f8d4e5c main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1784432 T3973 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1784432 T3973 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50157 collection:collection1 shard:shard1
   [junit4]   2> 1784432 T3973 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1784462 T3973 oascc.SolrZkClient.makePath makePath: /collections/collection1/leader_elect/shard1/election
   [junit4]   2> 1784514 T3973 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 1784535 T3973 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 1784536 T3973 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1784537 T3973 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50157/collection1/
   [junit4]   2> 1784537 T3973 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 1784537 T3973 oasc.SyncStrategy.syncToMe http://127.0.0.1:50157/collection1/ has no replicas
   [junit4]   2> 1784538 T3973 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50157/collection1/ shard1
   [junit4]   2> 1784538 T3973 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard1
   [junit4]   2> 1785235 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1785270 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1785270 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1785271 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1785326 T3973 oasc.ZkController.register We are http://127.0.0.1:50157/collection1/ and leader is http://127.0.0.1:50157/collection1/
   [junit4]   2> 1785328 T3973 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 1785328 T3973 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1785329 T3973 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1785333 T3973 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1785336 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1785337 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1785337 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1786184 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1786203 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50160
   [junit4]   2> 1786204 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1786204 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1786205 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545
   [junit4]   2> 1786206 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545/'
   [junit4]   2> 1786284 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545/solr.xml
   [junit4]   2> 1786481 T3930 oasc.CoreContainer.<init> New CoreContainer 1889282372
   [junit4]   2> 1786482 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545/]
   [junit4]   2> 1786485 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1786485 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1786486 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1786486 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1786487 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1786487 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1786487 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1786488 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1786488 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1786489 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1786519 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1786520 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1786521 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1786530 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1786535 T3986 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@34cb60eb name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1786538 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1786551 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1786576 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1786582 T3988 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@5b475631 name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1786582 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1786613 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1786802 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1786809 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50157_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50157"}
   [junit4]   2> 1786821 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1786822 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1786822 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1786825 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1787631 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50160_
   [junit4]   2> 1787638 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50160_
   [junit4]   2> 1787646 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1787650 T3988 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 1787650 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1787652 T3972 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 1787653 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1787654 T3952 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 1787654 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1787662 T3959 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 1787677 T3989 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1787678 T3989 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1787681 T3989 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 1788347 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1788351 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50160_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50160"}
   [junit4]   2> 1788352 T3953 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=4
   [junit4]   2> 1788352 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard4
   [junit4]   2> 1788365 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1788365 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1788365 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1788365 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1788683 T3989 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1788683 T3989 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545/collection1
   [junit4]   2> 1788684 T3989 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 1788688 T3989 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1788688 T3989 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 1788691 T3989 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545/collection1/'
   [junit4]   2> 1788693 T3989 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545/collection1/lib/classes/' to classloader
   [junit4]   2> 1788694 T3989 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545/collection1/lib/README' to classloader
   [junit4]   2> 1788801 T3989 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 1788929 T3989 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1788935 T3989 oass.IndexSchema.readSchema Reading Solr Schema from managed-schema
   [junit4]   2> 1788952 T3989 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1790966 T3989 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1790985 T3989 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1790998 T3989 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1791016 T3989 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1791040 T3989 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1791136 T3989 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1791154 T3989 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1791155 T3989 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1791156 T3989 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1791171 T3989 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1791171 T3989 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1791209 T3989 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1791210 T3989 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty2-1373827522545/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2/
   [junit4]   2> 1791210 T3989 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1791214 T3989 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2
   [junit4]   2> 1791214 T3989 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2/index/
   [junit4]   2> 1791215 T3989 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2/index' doesn't exist. Creating new index...
   [junit4]   2> 1791215 T3989 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2/index
   [junit4]   2> 1791222 T3989 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5aa906d1 lockFactory=org.apache.lucene.store.NativeFSLockFactory@6dc3105a),segFN=segments_1,generation=1}
   [junit4]   2> 1791222 T3989 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1791235 T3989 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1791239 T3989 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1791241 T3989 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1791241 T3989 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1791242 T3989 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1791250 T3989 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1791252 T3989 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1791270 T3989 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1791282 T3989 oass.SolrIndexSearcher.<init> Opening Searcher@48652d74 main
   [junit4]   2> 1791283 T3989 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1791284 T3989 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1791285 T3989 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1791301 T3990 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@48652d74 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1791307 T3989 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1791307 T3989 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50160 collection:collection1 shard:shard4
   [junit4]   2> 1791307 T3989 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1791356 T3989 oascc.SolrZkClient.makePath makePath: /collections/collection1/leader_elect/shard4/election
   [junit4]   2> 1791416 T3989 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 1791428 T3989 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 1791428 T3989 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1791428 T3989 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50160/collection1/
   [junit4]   2> 1791429 T3989 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 1791429 T3989 oasc.SyncStrategy.syncToMe http://127.0.0.1:50160/collection1/ has no replicas
   [junit4]   2> 1791429 T3989 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50160/collection1/ shard4
   [junit4]   2> 1791430 T3989 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard4
   [junit4]   2> 1792902 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1792938 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1792939 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1792946 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1792947 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1792971 T3989 oasc.ZkController.register We are http://127.0.0.1:50160/collection1/ and leader is http://127.0.0.1:50160/collection1/
   [junit4]   2> 1792972 T3989 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 1792973 T3989 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1792973 T3989 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1792980 T3989 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1792984 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1792985 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1792985 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1793964 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1793980 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50163
   [junit4]   2> 1793981 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1793981 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1793982 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197
   [junit4]   2> 1793982 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197/'
   [junit4]   2> 1794065 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197/solr.xml
   [junit4]   2> 1794217 T3930 oasc.CoreContainer.<init> New CoreContainer 502355220
   [junit4]   2> 1794217 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197/]
   [junit4]   2> 1794219 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1794220 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1794220 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1794221 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1794221 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1794223 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1794224 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1794224 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1794224 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1794225 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1794245 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1794246 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1794246 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1794248 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1794263 T4002 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@6ea68539 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1794263 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1794276 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1794317 T4004 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@14163885 name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1794318 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1794319 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1794359 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1794472 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1794476 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50160_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50160"}
   [junit4]   2> 1794497 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1794498 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1794498 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1794499 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1794506 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1795396 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50163_
   [junit4]   2> 1795407 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50163_
   [junit4]   2> 1795420 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1795423 T3988 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1795423 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1795424 T3952 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1795424 T4004 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1795425 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1795424 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1795424 T3972 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1795427 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1795433 T3959 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1795444 T4005 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1795445 T4005 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1795448 T4005 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 1796032 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1796037 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50163_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50163"}
   [junit4]   2> 1796038 T3953 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=4
   [junit4]   2> 1796038 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard2
   [junit4]   2> 1796060 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1796061 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1796061 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1796062 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1796061 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1796950 T4005 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1796951 T4005 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197/collection1
   [junit4]   2> 1796951 T4005 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 1796953 T4005 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1796953 T4005 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 1796955 T4005 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197/collection1/'
   [junit4]   2> 1796957 T4005 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197/collection1/lib/classes/' to classloader
   [junit4]   2> 1796958 T4005 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197/collection1/lib/README' to classloader
   [junit4]   2> 1797021 T4005 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 1797107 T4005 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1797110 T4005 oass.IndexSchema.readSchema Reading Solr Schema from managed-schema
   [junit4]   2> 1797122 T4005 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1798472 T4005 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1798502 T4005 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1798512 T4005 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1798525 T4005 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1798543 T4005 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1798587 T4005 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1798591 T4005 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1798591 T4005 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1798592 T4005 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1798596 T4005 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1798597 T4005 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1798611 T4005 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1798612 T4005 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty3-1373827530197/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3/
   [junit4]   2> 1798613 T4005 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1798623 T4005 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3
   [junit4]   2> 1798623 T4005 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3/index/
   [junit4]   2> 1798624 T4005 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3/index' doesn't exist. Creating new index...
   [junit4]   2> 1798625 T4005 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3/index
   [junit4]   2> 1798632 T4005 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@612c4ede lockFactory=org.apache.lucene.store.NativeFSLockFactory@daaade6),segFN=segments_1,generation=1}
   [junit4]   2> 1798633 T4005 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1798647 T4005 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1798649 T4005 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1798651 T4005 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1798652 T4005 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1798652 T4005 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1798654 T4005 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1798655 T4005 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1798690 T4005 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1798702 T4005 oass.SolrIndexSearcher.<init> Opening Searcher@3aa9f898 main
   [junit4]   2> 1798704 T4005 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1798705 T4005 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1798706 T4005 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1798719 T4006 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@3aa9f898 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1798721 T4005 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1798721 T4005 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50163 collection:collection1 shard:shard2
   [junit4]   2> 1798722 T4005 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1798756 T4005 oascc.SolrZkClient.makePath makePath: /collections/collection1/leader_elect/shard2/election
   [junit4]   2> 1798922 T4005 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 1798931 T4005 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 1798932 T4005 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1798932 T4005 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50163/collection1/
   [junit4]   2> 1798932 T4005 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 1798933 T4005 oasc.SyncStrategy.syncToMe http://127.0.0.1:50163/collection1/ has no replicas
   [junit4]   2> 1798933 T4005 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50163/collection1/ shard2
   [junit4]   2> 1798933 T4005 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard2
   [junit4]   2> 1799108 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1799131 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1799136 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1799137 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1799138 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1799138 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1799153 T4005 oasc.ZkController.register We are http://127.0.0.1:50163/collection1/ and leader is http://127.0.0.1:50163/collection1/
   [junit4]   2> 1799154 T4005 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 1799154 T4005 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1799155 T4005 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1799159 T4005 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1799164 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1799165 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1799166 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1800326 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1800732 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1800748 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50163_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50163"}
   [junit4]   2> 1800758 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50166
   [junit4]   2> 1800760 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1800760 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1800761 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393
   [junit4]   2> 1800761 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393/'
   [junit4]   2> 1800779 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1800779 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1800780 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1800780 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1800780 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1800842 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393/solr.xml
   [junit4]   2> 1800991 T3930 oasc.CoreContainer.<init> New CoreContainer 800665664
   [junit4]   2> 1800993 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393/]
   [junit4]   2> 1800995 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1800996 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1800997 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1800997 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1800998 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1800998 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1800998 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1800999 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1801000 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1801000 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1801029 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1801030 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1801030 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1801045 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1801051 T4018 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@23445a5 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1801052 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1801065 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1801088 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1801099 T4020 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@2419435 name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1801110 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1801138 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1802152 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50166_
   [junit4]   2> 1802206 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50166_
   [junit4]   2> 1802236 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1802245 T3952 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1802245 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1802278 T3972 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1802278 T4004 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1802279 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1802279 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1802284 T3988 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1802284 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1802286 T4020 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1802287 T3959 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1802324 T4021 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1802325 T4021 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1802327 T4021 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 1803805 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1803810 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50166_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50166"}
   [junit4]   2> 1803810 T3953 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=4
   [junit4]   2> 1803811 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard3
   [junit4]   2> 1803825 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1803825 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1803831 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1803832 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1803832 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1803833 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1804330 T4021 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1804330 T4021 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393/collection1
   [junit4]   2> 1804330 T4021 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 1804332 T4021 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1804332 T4021 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 1804340 T4021 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393/collection1/'
   [junit4]   2> 1804342 T4021 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393/collection1/lib/classes/' to classloader
   [junit4]   2> 1804343 T4021 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393/collection1/lib/README' to classloader
   [junit4]   2> 1804412 T4021 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 1804511 T4021 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1804515 T4021 oass.IndexSchema.readSchema Reading Solr Schema from managed-schema
   [junit4]   2> 1804543 T4021 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1805483 T4021 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1805494 T4021 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1805498 T4021 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1805505 T4021 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1805514 T4021 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1805547 T4021 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1805550 T4021 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1805550 T4021 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1805550 T4021 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1805553 T4021 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1805554 T4021 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1805562 T4021 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1805563 T4021 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty4-1373827536393/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4/
   [junit4]   2> 1805563 T4021 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1805584 T4021 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4
   [junit4]   2> 1805585 T4021 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4/index/
   [junit4]   2> 1805585 T4021 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4/index' doesn't exist. Creating new index...
   [junit4]   2> 1805586 T4021 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4/index
   [junit4]   2> 1805596 T4021 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@434e2c39 lockFactory=org.apache.lucene.store.NativeFSLockFactory@4293fe22),segFN=segments_1,generation=1}
   [junit4]   2> 1805598 T4021 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1805613 T4021 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1805617 T4021 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1805622 T4021 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1805623 T4021 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1805624 T4021 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1805625 T4021 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1805626 T4021 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1805670 T4021 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1805681 T4021 oass.SolrIndexSearcher.<init> Opening Searcher@346077bf main
   [junit4]   2> 1805684 T4021 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1805684 T4021 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1805686 T4021 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1805700 T4022 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@346077bf main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1805705 T4021 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1805705 T4021 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50166 collection:collection1 shard:shard3
   [junit4]   2> 1805706 T4021 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1805729 T4021 oascc.SolrZkClient.makePath makePath: /collections/collection1/leader_elect/shard3/election
   [junit4]   2> 1805746 T4021 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard3
   [junit4]   2> 1805753 T4021 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 1805753 T4021 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1805754 T4021 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50166/collection1/
   [junit4]   2> 1805754 T4021 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 1805755 T4021 oasc.SyncStrategy.syncToMe http://127.0.0.1:50166/collection1/ has no replicas
   [junit4]   2> 1805755 T4021 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50166/collection1/ shard3
   [junit4]   2> 1805755 T4021 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard3
   [junit4]   2> 1806851 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1806901 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1806902 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1806902 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1806907 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1806910 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1806914 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1806939 T4021 oasc.ZkController.register We are http://127.0.0.1:50166/collection1/ and leader is http://127.0.0.1:50166/collection1/
   [junit4]   2> 1806939 T4021 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 1806939 T4021 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1806940 T4021 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1806943 T4021 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1806946 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1806947 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1806948 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1807801 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1807824 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50169
   [junit4]   2> 1807827 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1807828 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1807829 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160
   [junit4]   2> 1807830 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160/'
   [junit4]   2> 1807899 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160/solr.xml
   [junit4]   2> 1808069 T3930 oasc.CoreContainer.<init> New CoreContainer 180980566
   [junit4]   2> 1808070 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160/]
   [junit4]   2> 1808072 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1808073 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1808074 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1808074 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1808075 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1808076 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1808076 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1808077 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1808077 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1808078 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1808130 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1808132 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1808133 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1808137 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1808144 T4034 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@3c597944 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1808145 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1808156 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1808190 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1808199 T4036 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@334f135d name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1808199 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1808212 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1808453 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1808463 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50166_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50166"}
   [junit4]   2> 1808484 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1808485 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1808485 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1808485 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1808492 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1808493 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1808494 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1809225 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50169_
   [junit4]   2> 1809229 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50169_
   [junit4]   2> 1809237 T3952 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1809237 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1809237 T3988 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1809237 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1809238 T3972 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1809238 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1809243 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1809247 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1809255 T4036 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1809256 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1809258 T4004 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1809259 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1809278 T3959 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1809280 T4020 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1809291 T4037 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1809291 T4037 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1809294 T4037 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 1810020 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1810022 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50169_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50169"}
   [junit4]   2> 1810023 T3953 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=4
   [junit4]   2> 1810023 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard1
   [junit4]   2> 1810098 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1810099 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1810100 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1810101 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1810101 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1810102 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1810104 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1810296 T4037 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1810297 T4037 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160/collection1
   [junit4]   2> 1810297 T4037 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 1810299 T4037 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1810300 T4037 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 1810303 T4037 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160/collection1/'
   [junit4]   2> 1810305 T4037 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160/collection1/lib/classes/' to classloader
   [junit4]   2> 1810306 T4037 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160/collection1/lib/README' to classloader
   [junit4]   2> 1810407 T4037 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 1810503 T4037 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1810506 T4037 oass.IndexSchema.readSchema Reading Solr Schema from managed-schema
   [junit4]   2> 1810520 T4037 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1811500 T4037 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1811514 T4037 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1811518 T4037 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1811527 T4037 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1811537 T4037 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1811567 T4037 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1811577 T4037 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1811578 T4037 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1811578 T4037 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1811582 T4037 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1811582 T4037 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1811591 T4037 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1811591 T4037 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty5-1373827544160/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5/
   [junit4]   2> 1811591 T4037 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1811593 T4037 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5
   [junit4]   2> 1811593 T4037 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5/index/
   [junit4]   2> 1811593 T4037 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5/index' doesn't exist. Creating new index...
   [junit4]   2> 1811594 T4037 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5/index
   [junit4]   2> 1811598 T4037 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@55c9c73a lockFactory=org.apache.lucene.store.NativeFSLockFactory@352d47e9),segFN=segments_1,generation=1}
   [junit4]   2> 1811599 T4037 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1811608 T4037 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1811609 T4037 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1811610 T4037 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1811610 T4037 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1811611 T4037 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1811611 T4037 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1811612 T4037 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1811639 T4037 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1811653 T4037 oass.SolrIndexSearcher.<init> Opening Searcher@7ece27a main
   [junit4]   2> 1811655 T4037 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1811656 T4037 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1811657 T4037 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1811670 T4037 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1811671 T4037 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50169 collection:collection1 shard:shard1
   [junit4]   2> 1811671 T4037 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1811677 T4038 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@7ece27a main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1811708 T4037 oasc.ZkController.register We are http://127.0.0.1:50169/collection1/ and leader is http://127.0.0.1:50157/collection1/
   [junit4]   2> 1811708 T4037 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:50169
   [junit4]   2> 1811709 T4037 oasc.ZkController.checkRecovery Core needs to recover:collection1
   [junit4]   2> 1811709 T4037 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> ASYNC  NEW_CORE C1040 name=collection1 org.apache.solr.core.SolrCore@67ead03 url=http://127.0.0.1:50169/collection1 node=127.0.0.1:50169_ C1040_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:50169_, base_url=http://127.0.0.1:50169}
   [junit4]   2> 1811710 T4039 C1040 P50169 oasc.RecoveryStrategy.run Starting recovery process.  core=collection1 recoveringAfterStartup=true
   [junit4]   2> 1811714 T4039 C1040 P50169 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 1811718 T4039 C1040 P50169 oasc.ZkController.publish publishing core=collection1 state=recovering
   [junit4]   2> 1811719 T4039 C1040 P50169 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1811719 T4037 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1811727 T4039 C1040 P50169 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1811740 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1811741 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1811742 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1811766 T3965 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node5, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 1812337 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1812373 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50173
   [junit4]   2> 1812374 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1812375 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1812375 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955
   [junit4]   2> 1812376 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955/'
   [junit4]   2> 1812428 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955/solr.xml
   [junit4]   2> 1812570 T3930 oasc.CoreContainer.<init> New CoreContainer 1192201794
   [junit4]   2> 1812571 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955/]
   [junit4]   2> 1812574 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1812575 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1812575 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1812576 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1812576 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1812577 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1812578 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1812578 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1812579 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1812580 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1812598 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1812599 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1812599 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1812605 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1812606 T4051 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@56a390d6 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1812607 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1812613 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1812621 T3932 oazs.NIOServerCnxn.doIO WARN caught end of stream exception EndOfStreamException: Unable to read additional data from client sessionid 0x13fde802b31000f, likely client has closed socket
   [junit4]   2> 	at org.apache.zookeeper.server.NIOServerCnxn.doIO(NIOServerCnxn.java:220)
   [junit4]   2> 	at org.apache.zookeeper.server.NIOServerCnxnFactory.run(NIOServerCnxnFactory.java:208)
   [junit4]   2> 	at java.lang.Thread.run(Thread.java:724)
   [junit4]   2> 
   [junit4]   2> 1812633 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1812637 T4053 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@5e93ba49 name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1812638 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1812651 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1813146 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1813150 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50169_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50169"}
   [junit4]   2> 1813160 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813168 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813169 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813169 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813169 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813170 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813170 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813174 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813665 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50173_
   [junit4]   2> 1813669 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50173_
   [junit4]   2> 1813674 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813678 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813678 T4004 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1813679 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1813679 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1813683 T3952 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1813683 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1813686 T3988 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1813686 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1813690 T4053 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1813695 T4036 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1813698 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1813701 T3972 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1813701 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1813706 T4020 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1813708 T3959 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1813715 T4054 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1813716 T4054 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1813718 T4054 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 1813772 T3965 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node5, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 1813773 T3965 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node5&state=recovering&nodeName=127.0.0.1:50169_&action=PREPRECOVERY&checkLive=true&core=collection1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2007 
   [junit4]   2> 1814692 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1814697 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50173_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50173"}
   [junit4]   2> 1814697 T3953 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=4
   [junit4]   2> 1814697 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard4
   [junit4]   2> 1814711 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1814712 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1814713 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1814713 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1814727 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1814736 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1814737 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1814737 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1815721 T4054 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1815721 T4054 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955/collection1
   [junit4]   2> 1815722 T4054 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 1815723 T4054 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1815724 T4054 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 1815726 T4054 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955/collection1/'
   [junit4]   2> 1815728 T4054 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955/collection1/lib/classes/' to classloader
   [junit4]   2> 1815729 T4054 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955/collection1/lib/README' to classloader
   [junit4]   2> ASYNC  NEW_CORE C1041 name=collection1 org.apache.solr.core.SolrCore@67ead03 url=http://127.0.0.1:50169/collection1 node=127.0.0.1:50169_ C1041_STATE=coll:collection1 core:collection1 props:{state=recovering, core=collection1, node_name=127.0.0.1:50169_, base_url=http://127.0.0.1:50169}
   [junit4]   2> 1815787 T4039 C1041 P50169 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50157/collection1/ core=collection1 - recoveringAfterStartup=true
   [junit4]   2> 1815789 T4039 C1041 P50169 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:50169 START replicas=[http://127.0.0.1:50157/collection1/] nUpdates=100
   [junit4]   2> 1815813 T4039 C1041 P50169 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 1815814 T4039 C1041 P50169 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=collection1
   [junit4]   2> 1815814 T4039 C1041 P50169 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=collection1
   [junit4]   2> 1815814 T4039 C1041 P50169 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=collection1
   [junit4]   2> 1815815 T4039 C1041 P50169 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 1815815 T4039 C1041 P50169 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50157/collection1/. core=collection1
   [junit4]   2> 1815815 T4039 C1041 P50169 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1815834 T4054 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> ASYNC  NEW_CORE C1042 name=collection1 org.apache.solr.core.SolrCore@796fdb0e url=http://127.0.0.1:50157/collection1 node=127.0.0.1:50157_ C1042_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50157_, base_url=http://127.0.0.1:50157, leader=true}
   [junit4]   2> 1815887 T3966 C1042 P50157 oasc.SolrCore.execute [collection1] webapp= path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=28 
   [junit4]   2> 1815929 T3966 C1042 P50157 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 1815944 T3966 C1042 P50157 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@64d807a9 lockFactory=org.apache.lucene.store.NativeFSLockFactory@4645c508),segFN=segments_1,generation=1}
   [junit4]   2> 1815946 T3966 C1042 P50157 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1815947 T3966 C1042 P50157 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 1815948 T3966 C1042 P50157 oass.SolrIndexSearcher.<init> Opening Searcher@31991f54 realtime
   [junit4]   2> 1815948 T3966 C1042 P50157 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 1815949 T3966 C1042 P50157 oasup.LogUpdateProcessor.finish [collection1] webapp= path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 20
   [junit4]   2> 1815951 T4039 C1041 P50169 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 1815951 T4039 C1041 P50169 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 1815978 T3964 C1042 P50157 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 1815978 T3964 C1042 P50157 oasc.SolrCore.execute [collection1] webapp= path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=2 
   [junit4]   2> 1815980 T4039 C1041 P50169 oasc.RecoveryStrategy.replay No replay needed. core=collection1
   [junit4]   2> 1815981 T4039 C1041 P50169 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=collection1
   [junit4]   2> 1815981 T4039 C1041 P50169 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1815981 T4039 C1041 P50169 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1815981 T4054 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1815989 T4039 C1041 P50169 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=collection1
   [junit4]   2> 1815995 T4054 oass.IndexSchema.readSchema Reading Solr Schema from managed-schema
   [junit4]   2> 1816009 T4054 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1816258 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1816263 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50169_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50169"}
   [junit4]   2> 1816283 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1816284 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1816285 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1816284 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1816283 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1816285 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1816284 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1816284 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1817290 T4054 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1817309 T4054 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1817316 T4054 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1817326 T4054 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1817340 T4054 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1817408 T4054 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1817417 T4054 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1817418 T4054 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1817418 T4054 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1817435 T4054 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1817439 T4054 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1817465 T4054 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1817466 T4054 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty6-1373827548955/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6/
   [junit4]   2> 1817467 T4054 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1817468 T4054 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6
   [junit4]   2> 1817468 T4054 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6/index/
   [junit4]   2> 1817469 T4054 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6/index' doesn't exist. Creating new index...
   [junit4]   2> 1817473 T4054 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6/index
   [junit4]   2> 1817481 T4054 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@76252f01 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7415efcd),segFN=segments_1,generation=1}
   [junit4]   2> 1817482 T4054 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1817492 T4054 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1817493 T4054 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1817494 T4054 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1817495 T4054 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1817498 T4054 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1817499 T4054 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1817506 T4054 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1817533 T4054 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1817543 T4054 oass.SolrIndexSearcher.<init> Opening Searcher@10e5d33 main
   [junit4]   2> 1817547 T4054 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1817548 T4054 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1817548 T4054 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1817562 T4056 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@10e5d33 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1817571 T4054 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1817572 T4054 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50173 collection:collection1 shard:shard4
   [junit4]   2> 1817576 T4054 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1817641 T4054 oasc.ZkController.register We are http://127.0.0.1:50173/collection1/ and leader is http://127.0.0.1:50160/collection1/
   [junit4]   2> 1817641 T4054 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:50173
   [junit4]   2> 1817642 T4054 oasc.ZkController.checkRecovery Core needs to recover:collection1
   [junit4]   2> 1817642 T4054 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> ASYNC  NEW_CORE C1043 name=collection1 org.apache.solr.core.SolrCore@35c1d950 url=http://127.0.0.1:50173/collection1 node=127.0.0.1:50173_ C1043_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:50173_, base_url=http://127.0.0.1:50173}
   [junit4]   2> 1817644 T4057 C1043 P50173 oasc.RecoveryStrategy.run Starting recovery process.  core=collection1 recoveringAfterStartup=true
   [junit4]   2> 1817645 T4057 C1043 P50173 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 1817645 T4057 C1043 P50173 oasc.ZkController.publish publishing core=collection1 state=recovering
   [junit4]   2> 1817646 T4057 C1043 P50173 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1817666 T4057 C1043 P50173 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1817673 T4054 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1817689 T3982 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node6, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 1817700 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1817704 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1817709 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1817807 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1817811 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50173_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50173"}
   [junit4]   2> 1818171 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1818171 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1818171 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1818172 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1818171 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1818171 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1818171 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1818171 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1818633 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1818663 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50180
   [junit4]   2> 1818664 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1818666 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1818666 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929
   [junit4]   2> 1818667 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929/'
   [junit4]   2> 1818691 T3982 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node6, state: recovering, checkLive: true, onlyIfLeader: true for: 1 seconds.
   [junit4]   2> 1818692 T3982 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node6&state=recovering&nodeName=127.0.0.1:50173_&action=PREPRECOVERY&checkLive=true&core=collection1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=1003 
   [junit4]   2> 1818739 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929/solr.xml
   [junit4]   2> 1818891 T3930 oasc.CoreContainer.<init> New CoreContainer 1765733726
   [junit4]   2> 1818892 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929/]
   [junit4]   2> 1818899 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1818901 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1818901 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1818902 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1818903 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1818904 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1818905 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1818905 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1818906 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1818907 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1818932 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1818934 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1818937 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1818946 T4069 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@78f162d6 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1818946 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1818947 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1818954 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1818983 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1818994 T4071 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@532efca name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1819001 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1819031 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1820053 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50180_
   [junit4]   2> 1820056 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50180_
   [junit4]   2> 1820062 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1820063 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1820064 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1820068 T3952 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820068 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1820069 T4004 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820069 T4036 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820069 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1820070 T3972 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820070 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1820069 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1820072 T3988 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820073 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1820071 T4071 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820079 T3959 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820080 T4053 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820080 T4020 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1820094 T4072 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1820094 T4072 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1820106 T4072 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2>  C1043_STATE=coll:collection1 core:collection1 props:{state=recovering, core=collection1, node_name=127.0.0.1:50173_, base_url=http://127.0.0.1:50173}
   [junit4]   2> 1820695 T4057 C1043 P50173 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50160/collection1/ core=collection1 - recoveringAfterStartup=true
   [junit4]   2> 1820696 T4057 C1043 P50173 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:50173 START replicas=[http://127.0.0.1:50160/collection1/] nUpdates=100
   [junit4]   2> 1820696 T4057 C1043 P50173 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 1820697 T4057 C1043 P50173 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=collection1
   [junit4]   2> 1820697 T4057 C1043 P50173 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=collection1
   [junit4]   2> 1820697 T4057 C1043 P50173 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=collection1
   [junit4]   2> 1820697 T4057 C1043 P50173 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 1820698 T4057 C1043 P50173 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50160/collection1/. core=collection1
   [junit4]   2> 1820698 T4057 C1043 P50173 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> ASYNC  NEW_CORE C1044 name=collection1 org.apache.solr.core.SolrCore@7526861e url=http://127.0.0.1:50160/collection1 node=127.0.0.1:50160_ C1044_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50160_, base_url=http://127.0.0.1:50160, leader=true}
   [junit4]   2> 1820720 T3979 C1044 P50160 oasc.SolrCore.execute [collection1] webapp= path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=2 
   [junit4]   2> 1820723 T3980 C1044 P50160 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 1820759 T3980 C1044 P50160 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5aa906d1 lockFactory=org.apache.lucene.store.NativeFSLockFactory@6dc3105a),segFN=segments_1,generation=1}
   [junit4]   2> 1820761 T3980 C1044 P50160 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1820761 T3980 C1044 P50160 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 1820762 T3980 C1044 P50160 oass.SolrIndexSearcher.<init> Opening Searcher@5d179538 realtime
   [junit4]   2> 1820762 T3980 C1044 P50160 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 1820762 T3980 C1044 P50160 oasup.LogUpdateProcessor.finish [collection1] webapp= path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 39
   [junit4]   2> 1820766 T4057 C1043 P50173 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 1820766 T4057 C1043 P50173 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 1820772 T3983 C1044 P50160 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 1820772 T3983 C1044 P50160 oasc.SolrCore.execute [collection1] webapp= path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=1 
   [junit4]   2> 1820775 T4057 C1043 P50173 oasc.RecoveryStrategy.replay No replay needed. core=collection1
   [junit4]   2> 1820775 T4057 C1043 P50173 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=collection1
   [junit4]   2> 1820776 T4057 C1043 P50173 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1820776 T4057 C1043 P50173 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1821009 T4057 C1043 P50173 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=collection1
   [junit4]   2> 1821197 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1821200 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50180_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50180"}
   [junit4]   2> 1821201 T3953 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=4
   [junit4]   2> 1821201 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard2
   [junit4]   2> 1821211 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50173_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50173"}
   [junit4]   2> 1821229 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1821229 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1821229 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1821230 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1821230 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1821229 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1821230 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1821230 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1821238 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1822108 T4072 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1822109 T4072 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929/collection1
   [junit4]   2> 1822109 T4072 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 1822111 T4072 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1822112 T4072 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 1822115 T4072 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929/collection1/'
   [junit4]   2> 1822117 T4072 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929/collection1/lib/classes/' to classloader
   [junit4]   2> 1822118 T4072 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929/collection1/lib/README' to classloader
   [junit4]   2> 1822224 T4072 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 1822316 T4072 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1822321 T4072 oass.IndexSchema.readSchema Reading Solr Schema from managed-schema
   [junit4]   2> 1822337 T4072 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1823604 T4072 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1823616 T4072 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1823621 T4072 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1823630 T4072 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1823639 T4072 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1823673 T4072 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1823678 T4072 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1823678 T4072 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1823678 T4072 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1823681 T4072 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1823681 T4072 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1823689 T4072 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1823690 T4072 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty7-1373827554929/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7/
   [junit4]   2> 1823690 T4072 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1823691 T4072 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7
   [junit4]   2> 1823691 T4072 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7/index/
   [junit4]   2> 1823692 T4072 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7/index' doesn't exist. Creating new index...
   [junit4]   2> 1823692 T4072 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7/index
   [junit4]   2> 1823696 T4072 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@16cce3fe lockFactory=org.apache.lucene.store.NativeFSLockFactory@48a84d04),segFN=segments_1,generation=1}
   [junit4]   2> 1823697 T4072 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1823704 T4072 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1823707 T4072 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1823708 T4072 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1823708 T4072 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1823709 T4072 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1823710 T4072 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1823710 T4072 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1823725 T4072 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1823733 T4072 oass.SolrIndexSearcher.<init> Opening Searcher@50a863f4 main
   [junit4]   2> 1823734 T4072 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1823735 T4072 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1823735 T4072 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1823749 T4074 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@50a863f4 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1823752 T4072 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1823752 T4072 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50180 collection:collection1 shard:shard2
   [junit4]   2> 1823753 T4072 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1823786 T4072 oasc.ZkController.register We are http://127.0.0.1:50180/collection1/ and leader is http://127.0.0.1:50163/collection1/
   [junit4]   2> 1823786 T4072 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:50180
   [junit4]   2> 1823786 T4072 oasc.ZkController.checkRecovery Core needs to recover:collection1
   [junit4]   2> 1823787 T4072 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> ASYNC  NEW_CORE C1045 name=collection1 org.apache.solr.core.SolrCore@7a0eeb56 url=http://127.0.0.1:50180/collection1 node=127.0.0.1:50180_ C1045_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:50180_, base_url=http://127.0.0.1:50180}
   [junit4]   2> 1823788 T4075 C1045 P50180 oasc.RecoveryStrategy.run Starting recovery process.  core=collection1 recoveringAfterStartup=true
   [junit4]   2> 1823789 T4075 C1045 P50180 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 1823790 T4075 C1045 P50180 oasc.ZkController.publish publishing core=collection1 state=recovering
   [junit4]   2> 1823790 T4075 C1045 P50180 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1823792 T4075 C1045 P50180 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1823796 T4072 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1823806 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1823807 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1823808 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1823817 T3996 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node7, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 1824270 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1824274 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50180_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50180"}
   [junit4]   2> 1824288 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824288 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824289 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824295 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824295 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824288 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824289 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824289 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824288 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1824605 T3930 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 1824615 T3930 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:50187
   [junit4]   2> 1824618 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 1824619 T3930 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 1824619 T3930 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019
   [junit4]   2> 1824620 T3930 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019/'
   [junit4]   2> 1824711 T3930 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019/solr.xml
   [junit4]   2> 1824818 T3996 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node7, state: recovering, checkLive: true, onlyIfLeader: true for: 1 seconds.
   [junit4]   2> 1824819 T3996 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node7&state=recovering&nodeName=127.0.0.1:50180_&action=PREPRECOVERY&checkLive=true&core=collection1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=1002 
   [junit4]   2> 1824856 T3930 oasc.CoreContainer.<init> New CoreContainer 614722938
   [junit4]   2> 1824856 T3930 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019/]
   [junit4]   2> 1824859 T3930 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 1824859 T3930 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 1824860 T3930 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 1824860 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 1824861 T3930 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 1824861 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 1824862 T3930 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 1824862 T3930 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 1824863 T3930 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 1824863 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 1824896 T3930 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 1824899 T3930 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:50149/solr
   [junit4]   2> 1824899 T3930 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 1824910 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1824915 T4087 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@4e16c8e8 name:ZooKeeperConnection Watcher:127.0.0.1:50149 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1824915 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1824947 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 1824974 T3930 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 1824981 T4089 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@275357ab name:ZooKeeperConnection Watcher:127.0.0.1:50149/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 1824981 T3930 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 1825018 T3930 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 1826051 T3930 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50187_
   [junit4]   2> 1826055 T3930 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:50187_
   [junit4]   2> 1826060 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1826062 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1826063 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1826063 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1826068 T3952 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826068 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1826069 T3972 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826070 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1826072 T3988 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826072 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1826070 T4004 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826073 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1826071 T4089 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826070 T4036 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826075 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1826079 T4053 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826080 T3959 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826083 T4071 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826084 T4020 oascc.ZkStateReader$3.process Updating live nodes... (9)
   [junit4]   2> 1826095 T4090 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1826096 T4090 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1826099 T4090 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2>  C1045_STATE=coll:collection1 core:collection1 props:{state=recovering, core=collection1, node_name=127.0.0.1:50180_, base_url=http://127.0.0.1:50180}
   [junit4]   2> 1826822 T4075 C1045 P50180 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50163/collection1/ core=collection1 - recoveringAfterStartup=true
   [junit4]   2> 1826823 T4075 C1045 P50180 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:50180 START replicas=[http://127.0.0.1:50163/collection1/] nUpdates=100
   [junit4]   2> 1826824 T4075 C1045 P50180 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 1826824 T4075 C1045 P50180 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=collection1
   [junit4]   2> 1826825 T4075 C1045 P50180 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=collection1
   [junit4]   2> 1826825 T4075 C1045 P50180 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=collection1
   [junit4]   2> 1826826 T4075 C1045 P50180 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 1826826 T4075 C1045 P50180 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50163/collection1/. core=collection1
   [junit4]   2> 1826827 T4075 C1045 P50180 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> ASYNC  NEW_CORE C1046 name=collection1 org.apache.solr.core.SolrCore@2c20d3f7 url=http://127.0.0.1:50163/collection1 node=127.0.0.1:50163_ C1046_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50163_, base_url=http://127.0.0.1:50163, leader=true}
   [junit4]   2> 1826866 T3999 C1046 P50163 oasc.SolrCore.execute [collection1] webapp= path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=1 
   [junit4]   2> 1826893 T3995 C1046 P50163 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 1826900 T3995 C1046 P50163 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@612c4ede lockFactory=org.apache.lucene.store.NativeFSLockFactory@daaade6),segFN=segments_1,generation=1}
   [junit4]   2> 1826902 T3995 C1046 P50163 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1826905 T3995 C1046 P50163 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 1826908 T3995 C1046 P50163 oass.SolrIndexSearcher.<init> Opening Searcher@96383cb realtime
   [junit4]   2> 1826909 T3995 C1046 P50163 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 1826910 T3995 C1046 P50163 oasup.LogUpdateProcessor.finish [collection1] webapp= path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 18
   [junit4]   2> 1826912 T4075 C1045 P50180 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 1826913 T4075 C1045 P50180 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 1826921 T3998 C1046 P50163 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 1826923 T3998 C1046 P50163 oasc.SolrCore.execute [collection1] webapp= path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=5 
   [junit4]   2> 1826926 T4075 C1045 P50180 oasc.RecoveryStrategy.replay No replay needed. core=collection1
   [junit4]   2> 1826927 T4075 C1045 P50180 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=collection1
   [junit4]   2> 1826927 T4075 C1045 P50180 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1826927 T4075 C1045 P50180 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1826934 T4075 C1045 P50180 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=collection1
   [junit4]   2> 1827324 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1827326 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50187_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50187"}
   [junit4]   2> 1827326 T3953 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=4
   [junit4]   2> 1827327 T3953 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard3
   [junit4]   2> 1827334 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50180_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50180"}
   [junit4]   2> 1827343 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827343 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827344 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827347 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827347 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827347 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827348 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827349 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827349 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1827351 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1828109 T4090 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 1828109 T4090 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019/collection1
   [junit4]   2> 1828110 T4090 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 1828112 T4090 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 1828112 T4090 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 1828118 T4090 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019/collection1/'
   [junit4]   2> 1828120 T4090 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019/collection1/lib/classes/' to classloader
   [junit4]   2> 1828121 T4090 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019/collection1/lib/README' to classloader
   [junit4]   2> 1828249 T4090 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 1828420 T4090 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 1828425 T4090 oass.IndexSchema.readSchema Reading Solr Schema from managed-schema
   [junit4]   2> 1828451 T4090 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 1829761 T4090 oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 1829776 T4090 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 1829781 T4090 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 1829789 T4090 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1829797 T4090 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 1829822 T4090 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1829826 T4090 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1829827 T4090 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1829829 T4090 oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 1829837 T4090 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 1829837 T4090 oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 1829847 T4090 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 1829847 T4090 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.schema.TestCloudManagedSchemaAddField-jetty8-1373827561019/collection1/, dataDir=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8/
   [junit4]   2> 1829848 T4090 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 1829849 T4090 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8
   [junit4]   2> 1829850 T4090 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8/index/
   [junit4]   2> 1829850 T4090 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8/index' doesn't exist. Creating new index...
   [junit4]   2> 1829851 T4090 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8/index
   [junit4]   2> 1829856 T4090 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@ff66dda lockFactory=org.apache.lucene.store.NativeFSLockFactory@6c97500c),segFN=segments_1,generation=1}
   [junit4]   2> 1829857 T4090 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1829867 T4090 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 1829868 T4090 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 1829869 T4090 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 1829869 T4090 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 1829871 T4090 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 1829872 T4090 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 1829872 T4090 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 1829887 T4090 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 1829900 T4090 oass.SolrIndexSearcher.<init> Opening Searcher@422c5aec main
   [junit4]   2> 1829904 T4090 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 1829905 T4090 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 1829906 T4090 oass.ZkIndexSchemaReader.createSchemaWatcher Creating ZooKeeper watch for the managed schema at /configs/conf1/managed-schema ...
   [junit4]   2> 1829914 T4092 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@422c5aec main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 1829920 T4090 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 1829921 T4090 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50187 collection:collection1 shard:shard3
   [junit4]   2> 1829921 T4090 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 1829956 T4090 oasc.ZkController.register We are http://127.0.0.1:50187/collection1/ and leader is http://127.0.0.1:50166/collection1/
   [junit4]   2> 1829956 T4090 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:50187
   [junit4]   2> 1829957 T4090 oasc.ZkController.checkRecovery Core needs to recover:collection1
   [junit4]   2> 1829957 T4090 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> ASYNC  NEW_CORE C1047 name=collection1 org.apache.solr.core.SolrCore@66b4c55d url=http://127.0.0.1:50187/collection1 node=127.0.0.1:50187_ C1047_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:50187_, base_url=http://127.0.0.1:50187}
   [junit4]   2> 1829959 T4093 C1047 P50187 oasc.RecoveryStrategy.run Starting recovery process.  core=collection1 recoveringAfterStartup=true
   [junit4]   2> 1829969 T4093 C1047 P50187 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 1829970 T4093 C1047 P50187 oasc.ZkController.publish publishing core=collection1 state=recovering
   [junit4]   2> 1829970 T4093 C1047 P50187 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1829970 T4090 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1829972 T4093 C1047 P50187 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1829974 T3930 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 1829978 T3930 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 1829980 T3930 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1830020 T3930 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1830031 T4009 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node8, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> ASYNC  NEW_CORE C1048 name=collection1 org.apache.solr.core.SolrCore@67ead03 url=http://127.0.0.1:50169/collection1 node=127.0.0.1:50169_ C1048_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50169_, base_url=http://127.0.0.1:50169}
   [junit4]   2> 1830067 T4032 C1048 P50169 oejsh.ContextHandler$Context.log SolrRestApi: [Restlet] Attaching application: org.apache.solr.rest.SolrRestApi@36d893b6 to URI: /schema
   [junit4]   2> 1830079 T4032 C1048 P50169 oasr.SolrRestApi.createInboundRoot createInboundRoot started
   [junit4]   2> 1830082 T4032 C1048 P50169 oasr.SolrRestApi.createInboundRoot createInboundRoot complete
   [junit4]   2> 1830105 T4004 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830105 T3952 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830106 T4053 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830106 T4036 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830106 T3988 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830106 T4089 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830106 T3972 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830106 T4020 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830106 T4032 C1048 P50169 oass.ManagedIndexSchema.persistManagedSchemaToZooKeeper Persisted managed schema at /configs/conf1/managed-schema
   [junit4]   2> 1830109 T4032 C1048 P50169 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=47 
   [junit4]   2> 1830107 T4071 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> ASYNC  NEW_CORE C17 name=collection1 org.apache.solr.core.SolrCore@67ead03 url=http://127.0.0.1:50169/collection1 node=127.0.0.1:50169_ C17_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50169_, base_url=http://127.0.0.1:50169}
   [junit4]   2> 58330 T4032 C17 P50169 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50169	PUT	/schema/fields/newfield1	wt=xml	200	-	32	40	http://127.0.0.1:50169	Java/1.7.0_25	-
   [junit4]   2> 1830112 T4053 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830115 T4004 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830122 T3988 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830133 T3952 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830165 T4089 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830182 T4020 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830188 T3972 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830188 T4071 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830226 T4036 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830292 T3952 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 158 ms
   [junit4]   2> ASYNC  NEW_CORE C1049 name=collection1 org.apache.solr.core.SolrCore@796fdb0e url=http://127.0.0.1:50157/collection1 node=127.0.0.1:50157_ C1049_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50157_, base_url=http://127.0.0.1:50157, leader=true}
   [junit4]   2> 1830286 T3966 C1049 P50157 oejsh.ContextHandler$Context.log SolrRestApi: [Restlet] Attaching application: org.apache.solr.rest.SolrRestApi@6aca477a to URI: /schema
   [junit4]   2> 1830321 T3966 C1049 P50157 oasr.SolrRestApi.createInboundRoot createInboundRoot started
   [junit4]   2> 1830321 T4036 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 95 ms
   [junit4]   2> 1830322 T3966 C1049 P50157 oasr.SolrRestApi.createInboundRoot createInboundRoot complete
   [junit4]   2> 1830324 T3966 C1049 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} msg={Field 'newfield1' not found.} status=404 QTime=99 
   [junit4]   2> ASYNC  NEW_CORE C18 name=collection1 org.apache.solr.core.SolrCore@796fdb0e url=http://127.0.0.1:50157/collection1 node=127.0.0.1:50157_ C18_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50157_, base_url=http://127.0.0.1:50157, leader=true}
   [junit4]   2> 58544 T3966 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield1	wt=xml	404	-	0	6	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1830330 T4004 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 215 ms
   [junit4]   2> 1830334 T3988 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 211 ms
   [junit4]   2> 1830334 T3972 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 120 ms
   [junit4]   2> 1830338 T4089 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 172 ms
   [junit4]   2> 1830342 T4020 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 145 ms
   [junit4]   2> 1830342 T4071 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 127 ms
   [junit4]   2> 1830348 T4053 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 236 ms
   [junit4]   2> 1830377 T3964 C1049 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=4 
   [junit4]   2> 58598 T3964 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield1	wt=xml	200	-	0	3	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1830385 T3930 oass.TestCloudManagedSchemaAddField.doTest On attempt #2, successful request /schema/fields/newfield1?wt=xml against server http://127.0.0.1:50157/collection1 after 235 ms
   [junit4]   2> ASYNC  NEW_CORE C1050 name=collection1 org.apache.solr.core.SolrCore@7526861e url=http://127.0.0.1:50160/collection1 node=127.0.0.1:50160_ C1050_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50160_, base_url=http://127.0.0.1:50160, leader=true}
   [junit4]   2> 1830401 T3981 C1050 P50160 oejsh.ContextHandler$Context.log SolrRestApi: [Restlet] Attaching application: org.apache.solr.rest.SolrRestApi@d2af28b to URI: /schema
   [junit4]   2> 1830404 T3981 C1050 P50160 oasr.SolrRestApi.createInboundRoot createInboundRoot started
   [junit4]   2> 1830405 T3981 C1050 P50160 oasr.SolrRestApi.createInboundRoot createInboundRoot complete
   [junit4]   2> 1830407 T3981 C1050 P50160 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=19 
   [junit4]   2> ASYNC  NEW_CORE C19 name=collection1 org.apache.solr.core.SolrCore@7526861e url=http://127.0.0.1:50160/collection1 node=127.0.0.1:50160_ C19_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50160_, base_url=http://127.0.0.1:50160, leader=true}
   [junit4]   2> 58627 T3981 C19 P50160 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50160	GET	/schema/fields/newfield1	wt=xml	200	-	0	6	http://127.0.0.1:50160	Java/1.7.0_25	-
   [junit4]   2> 1830414 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1830417 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50187_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50187"}
   [junit4]   2> ASYNC  NEW_CORE C1051 name=collection1 org.apache.solr.core.SolrCore@2c20d3f7 url=http://127.0.0.1:50163/collection1 node=127.0.0.1:50163_ C1051_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50163_, base_url=http://127.0.0.1:50163, leader=true}
   [junit4]   2> 1830433 T3995 C1051 P50163 oejsh.ContextHandler$Context.log SolrRestApi: [Restlet] Attaching application: org.apache.solr.rest.SolrRestApi@4e95be63 to URI: /schema
   [junit4]   2> 1830438 T3995 C1051 P50163 oasr.SolrRestApi.createInboundRoot createInboundRoot started
   [junit4]   2> 1830441 T3959 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830443 T3995 C1051 P50163 oasr.SolrRestApi.createInboundRoot createInboundRoot complete
   [junit4]   2> 1830444 T3995 C1051 P50163 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=18 
   [junit4]   2> ASYNC  NEW_CORE C20 name=collection1 org.apache.solr.core.SolrCore@2c20d3f7 url=http://127.0.0.1:50163/collection1 node=127.0.0.1:50163_ C20_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50163_, base_url=http://127.0.0.1:50163, leader=true}
   [junit4]   2> 58667 T3995 C20 P50163 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50163	GET	/schema/fields/newfield1	wt=xml	200	-	0	13	http://127.0.0.1:50163	Java/1.7.0_25	-
   [junit4]   2> 1830449 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830450 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830452 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830452 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830456 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830453 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830466 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830475 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> ASYNC  NEW_CORE C1052 name=collection1 org.apache.solr.core.SolrCore@4300ec6e url=http://127.0.0.1:50166/collection1 node=127.0.0.1:50166_ C1052_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50166_, base_url=http://127.0.0.1:50166, leader=true}
   [junit4]   2> 1830482 T4013 C1052 P50166 oejsh.ContextHandler$Context.log SolrRestApi: [Restlet] Attaching application: org.apache.solr.rest.SolrRestApi@363f3a90 to URI: /schema
   [junit4]   2> 1830486 T4013 C1052 P50166 oasr.SolrRestApi.createInboundRoot createInboundRoot started
   [junit4]   2> 1830487 T4013 C1052 P50166 oasr.SolrRestApi.createInboundRoot createInboundRoot complete
   [junit4]   2> 1830488 T4013 C1052 P50166 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=16 
   [junit4]   2> ASYNC  NEW_CORE C21 name=collection1 org.apache.solr.core.SolrCore@4300ec6e url=http://127.0.0.1:50166/collection1 node=127.0.0.1:50166_ C21_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50166_, base_url=http://127.0.0.1:50166, leader=true}
   [junit4]   2> 58709 T4013 C21 P50166 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50166	GET	/schema/fields/newfield1	wt=xml	200	-	0	7	http://127.0.0.1:50166	Java/1.7.0_25	-
   [junit4]   2> 1830490 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1830500 T4027 C1048 P50169 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=3 
   [junit4]   2> 58720 T4027 C17 P50169 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50169	GET	/schema/fields/newfield1	wt=xml	200	-	0	2	http://127.0.0.1:50169	Java/1.7.0_25	-
   [junit4]   2> ASYNC  NEW_CORE C1053 name=collection1 org.apache.solr.core.SolrCore@35c1d950 url=http://127.0.0.1:50173/collection1 node=127.0.0.1:50173_ C1053_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50173_, base_url=http://127.0.0.1:50173}
   [junit4]   2> 1830520 T4049 C1053 P50173 oejsh.ContextHandler$Context.log SolrRestApi: [Restlet] Attaching application: org.apache.solr.rest.SolrRestApi@77ec430f to URI: /schema
   [junit4]   2> 1830524 T4049 C1053 P50173 oasr.SolrRestApi.createInboundRoot createInboundRoot started
   [junit4]   2> 1830525 T4049 C1053 P50173 oasr.SolrRestApi.createInboundRoot createInboundRoot complete
   [junit4]   2> 1830536 T4049 C1053 P50173 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=18 
   [junit4]   2> ASYNC  NEW_CORE C22 name=collection1 org.apache.solr.core.SolrCore@35c1d950 url=http://127.0.0.1:50173/collection1 node=127.0.0.1:50173_ C22_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50173_, base_url=http://127.0.0.1:50173}
   [junit4]   2> 58756 T4049 C22 P50173 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50173	GET	/schema/fields/newfield1	wt=xml	200	-	0	15	http://127.0.0.1:50173	Java/1.7.0_25	-
   [junit4]   2> ASYNC  NEW_CORE C1054 name=collection1 org.apache.solr.core.SolrCore@7a0eeb56 url=http://127.0.0.1:50180/collection1 node=127.0.0.1:50180_ C1054_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50180_, base_url=http://127.0.0.1:50180}
   [junit4]   2> 1830559 T4065 C1054 P50180 oejsh.ContextHandler$Context.log SolrRestApi: [Restlet] Attaching application: org.apache.solr.rest.SolrRestApi@4a8309a to URI: /schema
   [junit4]   2> 1830561 T4065 C1054 P50180 oasr.SolrRestApi.createInboundRoot createInboundRoot started
   [junit4]   2> 1830563 T4065 C1054 P50180 oasr.SolrRestApi.createInboundRoot createInboundRoot complete
   [junit4]   2> 1830564 T4065 C1054 P50180 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=10 
   [junit4]   2> ASYNC  NEW_CORE C23 name=collection1 org.apache.solr.core.SolrCore@7a0eeb56 url=http://127.0.0.1:50180/collection1 node=127.0.0.1:50180_ C23_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50180_, base_url=http://127.0.0.1:50180}
   [junit4]   2> 58785 T4065 C23 P50180 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50180	GET	/schema/fields/newfield1	wt=xml	200	-	0	6	http://127.0.0.1:50180	Java/1.7.0_25	-
   [junit4]   2>  C1047_STATE=coll:collection1 core:collection1 props:{state=recovering, core=collection1, node_name=127.0.0.1:50187_, base_url=http://127.0.0.1:50187}
   [junit4]   2> 1830592 T4082 C1047 P50187 oejsh.ContextHandler$Context.log SolrRestApi: [Restlet] Attaching application: org.apache.solr.rest.SolrRestApi@4449bd44 to URI: /schema
   [junit4]   2> 1830595 T4082 C1047 P50187 oasr.SolrRestApi.createInboundRoot createInboundRoot started
   [junit4]   2> 1830596 T4082 C1047 P50187 oasr.SolrRestApi.createInboundRoot createInboundRoot complete
   [junit4]   2> 1830598 T4082 C1047 P50187 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield1 params={wt=xml} status=0 QTime=8 
   [junit4]   2> ASYNC  NEW_CORE C24 name=collection1 org.apache.solr.core.SolrCore@66b4c55d url=http://127.0.0.1:50187/collection1 node=127.0.0.1:50187_ C24_STATE=coll:collection1 core:collection1 props:{state=recovering, core=collection1, node_name=127.0.0.1:50187_, base_url=http://127.0.0.1:50187}
   [junit4]   2> 58817 T4082 C24 P50187 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50187	GET	/schema/fields/newfield1	wt=xml	200	-	0	4	http://127.0.0.1:50187	Java/1.7.0_25	-
   [junit4]   2> 1830638 T4004 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830641 T3952 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830642 T3988 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830640 T4036 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830642 T4020 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830646 T4004 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830649 T3988 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830655 T4020 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830642 T3972 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830657 T3952 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830641 T4071 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830641 T4089 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830641 T4030 C1048 P50169 oass.ManagedIndexSchema.persistManagedSchemaToZooKeeper Persisted managed schema at /configs/conf1/managed-schema
   [junit4]   2> 1830661 T3972 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830663 T4071 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830682 T3988 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 33 ms
   [junit4]   2> 1830640 T4053 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830666 T4089 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830661 T4030 C1048 P50169 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=54 
   [junit4]   2> 58903 T4030 C17 P50169 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50169	PUT	/schema/fields/newfield2	wt=xml	200	-	32	77	http://127.0.0.1:50169	Java/1.7.0_25	-
   [junit4]   2> 1830692 T4071 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 19 ms
   [junit4]   2> 1830696 T4036 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830698 T4053 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830702 T3972 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 40 ms
   [junit4]   2> 1830715 T4053 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 17 ms
   [junit4]   2> 1830757 T3952 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 99 ms
   [junit4]   2> 1830759 T4089 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 76 ms
   [junit4]   2> 1830761 T4020 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 105 ms
   [junit4]   2> 1830763 T4004 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 117 ms
   [junit4]   2> 1830766 T4036 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 70 ms
   [junit4]   2> 1830766 T3963 C1049 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 58987 T3963 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield2	wt=xml	200	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1830789 T3983 C1050 P50160 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=16 
   [junit4]   2> 59009 T3983 C19 P50160 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50160	GET	/schema/fields/newfield2	wt=xml	200	-	0	3	http://127.0.0.1:50160	Java/1.7.0_25	-
   [junit4]   2> 1830801 T3997 C1051 P50163 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=6 
   [junit4]   2> 59023 T3997 C20 P50163 orel.LogFilter.afterHandle 2013-07-15	05:46:07	127.0.0.1	-	127.0.0.1	50163	GET	/schema/fields/newfield2	wt=xml	200	-	0	2	http://127.0.0.1:50163	Java/1.7.0_25	-
   [junit4]   2> 1830814 T4014 C1052 P50166 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=3 
   [junit4]   2> 59034 T4014 C21 P50166 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50166	GET	/schema/fields/newfield2	wt=xml	200	-	0	3	http://127.0.0.1:50166	Java/1.7.0_25	-
   [junit4]   2> 1830823 T4031 C1048 P50169 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59043 T4031 C17 P50169 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50169	GET	/schema/fields/newfield2	wt=xml	200	-	0	2	http://127.0.0.1:50169	Java/1.7.0_25	-
   [junit4]   2> 1830834 T4045 C1053 P50173 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=2 
   [junit4]   2> 59054 T4045 C22 P50173 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50173	GET	/schema/fields/newfield2	wt=xml	200	-	0	2	http://127.0.0.1:50173	Java/1.7.0_25	-
   [junit4]   2> 1830844 T4066 C1054 P50180 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=2 
   [junit4]   2> 59064 T4066 C23 P50180 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50180	GET	/schema/fields/newfield2	wt=xml	200	-	0	2	http://127.0.0.1:50180	Java/1.7.0_25	-
   [junit4]   2> 1830868 T4083 C1047 P50187 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield2 params={wt=xml} status=0 QTime=2 
   [junit4]   2> 59088 T4083 C24 P50187 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50187	GET	/schema/fields/newfield2	wt=xml	200	-	0	2	http://127.0.0.1:50187	Java/1.7.0_25	-
   [junit4]   2> 1830891 T4032 C1048 P50169 oass.ManagedIndexSchema.persistManagedSchemaToZooKeeper Persisted managed schema at /configs/conf1/managed-schema
   [junit4]   2> 1830891 T4032 C1048 P50169 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=10 
   [junit4]   2> 59111 T4032 C17 P50169 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50169	PUT	/schema/fields/newfield3	wt=xml	200	-	32	10	http://127.0.0.1:50169	Java/1.7.0_25	-
   [junit4]   2> 1830894 T4036 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830896 T4036 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830906 T3988 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830907 T4053 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830909 T4020 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830910 T4071 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830906 T4004 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830915 T4036 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 19 ms
   [junit4]   2> 1830916 T3988 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830917 T4089 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830918 T3952 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830919 T3972 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1830925 T4053 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830934 T4004 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830945 T4020 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830945 T3952 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830946 T4071 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830946 T3972 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830947 T4089 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1830955 T4053 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 30 ms
   [junit4]   2> 1830974 T3988 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 58 ms
   [junit4]   2> 1830979 T3967 C1049 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} msg={Field 'newfield3' not found.} status=404 QTime=1 
   [junit4]   2> 59199 T3967 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield3	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831009 T4020 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 64 ms
   [junit4]   2> 1831012 T4004 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 76 ms
   [junit4]   2> 1831017 T4071 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 71 ms
   [junit4]   2> 1831024 T3952 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 78 ms
   [junit4]   2> 1831031 T4089 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 84 ms
   [junit4]   2> 1831032 T4009 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node8, state: recovering, checkLive: true, onlyIfLeader: true for: 1 seconds.
   [junit4]   2> 1831033 T4009 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node8&state=recovering&nodeName=127.0.0.1:50187_&action=PREPRECOVERY&checkLive=true&core=collection1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=1003 
   [junit4]   2> 1831034 T3968 C1049 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} msg={Field 'newfield3' not found.} status=404 QTime=1 
   [junit4]   2> 59254 T3968 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield3	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831050 T3972 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 103 ms
   [junit4]   2> 1831065 T3966 C1049 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=2 
   [junit4]   2> 59286 T3966 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield3	wt=xml	200	-	0	4	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831075 T3930 oass.TestCloudManagedSchemaAddField.doTest On attempt #3, successful request /schema/fields/newfield3?wt=xml against server http://127.0.0.1:50157/collection1 after 94 ms
   [junit4]   2> 1831080 T3982 C1050 P50160 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=3 
   [junit4]   2> 59300 T3982 C19 P50160 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50160	GET	/schema/fields/newfield3	wt=xml	200	-	0	3	http://127.0.0.1:50160	Java/1.7.0_25	-
   [junit4]   2> 1831089 T3996 C1051 P50163 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59310 T3996 C20 P50163 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50163	GET	/schema/fields/newfield3	wt=xml	200	-	0	1	http://127.0.0.1:50163	Java/1.7.0_25	-
   [junit4]   2> 1831099 T4012 C1052 P50166 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=2 
   [junit4]   2> 59319 T4012 C21 P50166 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50166	GET	/schema/fields/newfield3	wt=xml	200	-	0	2	http://127.0.0.1:50166	Java/1.7.0_25	-
   [junit4]   2> 1831117 T4028 C1048 P50169 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=2 
   [junit4]   2> 59337 T4028 C17 P50169 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50169	GET	/schema/fields/newfield3	wt=xml	200	-	0	2	http://127.0.0.1:50169	Java/1.7.0_25	-
   [junit4]   2> 1831126 T4046 C1053 P50173 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59348 T4046 C22 P50173 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50173	GET	/schema/fields/newfield3	wt=xml	200	-	0	4	http://127.0.0.1:50173	Java/1.7.0_25	-
   [junit4]   2> 1831140 T4062 C1054 P50180 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59361 T4062 C23 P50180 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50180	GET	/schema/fields/newfield3	wt=xml	200	-	0	2	http://127.0.0.1:50180	Java/1.7.0_25	-
   [junit4]   2> 1831150 T4085 C1047 P50187 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield3 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59371 T4085 C24 P50187 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50187	GET	/schema/fields/newfield3	wt=xml	200	-	0	1	http://127.0.0.1:50187	Java/1.7.0_25	-
   [junit4]   2> 1831174 T4047 C1053 P50173 oass.ManagedIndexSchema.persistManagedSchemaToZooKeeper Persisted managed schema at /configs/conf1/managed-schema
   [junit4]   2> 1831175 T4036 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831175 T3972 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831175 T3952 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831174 T4071 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831174 T4089 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831180 T4071 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831174 T4004 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831174 T4053 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831181 T4036 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831183 T4053 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831201 T4053 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 18 ms
   [junit4]   2> 1831205 T4004 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831181 T3972 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831181 T3952 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831180 T4089 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831175 T3988 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831175 T4020 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> ASYNC  NEW_CORE C1055 name=collection1 org.apache.solr.core.SolrCore@35c1d950 url=http://127.0.0.1:50173/collection1 node=127.0.0.1:50173_ C1055_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50173_, base_url=http://127.0.0.1:50173}
   [junit4]   2> 1831175 T4047 C1055 P50173 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=11 
   [junit4]   2> 1831240 T4036 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 58 ms
   [junit4]   2> 1831266 T3988 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 59485 T4047 C22 P50173 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50173	PUT	/schema/fields/newfield4	wt=xml	200	-	32	100	http://127.0.0.1:50173	Java/1.7.0_25	-
   [junit4]   2> 1831277 T4020 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831280 T3952 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 38 ms
   [junit4]   2> 1831288 T3988 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 22 ms
   [junit4]   2> 1831306 T4071 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 125 ms
   [junit4]   2> 1831307 T4004 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 102 ms
   [junit4]   2> 1831316 T4089 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 64 ms
   [junit4]   2> 1831329 T4020 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 52 ms
   [junit4]   2> ASYNC  NEW_CORE C1056 name=collection1 org.apache.solr.core.SolrCore@796fdb0e url=http://127.0.0.1:50157/collection1 node=127.0.0.1:50157_ C1056_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50157_, base_url=http://127.0.0.1:50157, leader=true}
   [junit4]   2> 1831335 T3964 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} msg={Field 'newfield4' not found.} status=404 QTime=1 
   [junit4]   2> 59555 T3964 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield4	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831355 T3972 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 114 ms
   [junit4]   2> 1831364 T3963 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=2 
   [junit4]   2> 59584 T3963 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield4	wt=xml	200	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831371 T3930 oass.TestCloudManagedSchemaAddField.doTest On attempt #2, successful request /schema/fields/newfield4?wt=xml against server http://127.0.0.1:50157/collection1 after 37 ms
   [junit4]   2> ASYNC  NEW_CORE C1057 name=collection1 org.apache.solr.core.SolrCore@7526861e url=http://127.0.0.1:50160/collection1 node=127.0.0.1:50160_ C1057_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50160_, base_url=http://127.0.0.1:50160, leader=true}
   [junit4]   2> 1831384 T3979 C1057 P50160 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59604 T3979 C19 P50160 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50160	GET	/schema/fields/newfield4	wt=xml	200	-	0	2	http://127.0.0.1:50160	Java/1.7.0_25	-
   [junit4]   2> ASYNC  NEW_CORE C1058 name=collection1 org.apache.solr.core.SolrCore@2c20d3f7 url=http://127.0.0.1:50163/collection1 node=127.0.0.1:50163_ C1058_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50163_, base_url=http://127.0.0.1:50163, leader=true}
   [junit4]   2> 1831393 T3999 C1058 P50163 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59614 T3999 C20 P50163 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50163	GET	/schema/fields/newfield4	wt=xml	200	-	0	2	http://127.0.0.1:50163	Java/1.7.0_25	-
   [junit4]   2> ASYNC  NEW_CORE C1059 name=collection1 org.apache.solr.core.SolrCore@4300ec6e url=http://127.0.0.1:50166/collection1 node=127.0.0.1:50166_ C1059_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50166_, base_url=http://127.0.0.1:50166, leader=true}
   [junit4]   2> 1831414 T4013 C1059 P50166 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59634 T4013 C21 P50166 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50166	GET	/schema/fields/newfield4	wt=xml	200	-	0	1	http://127.0.0.1:50166	Java/1.7.0_25	-
   [junit4]   2> ASYNC  NEW_CORE C1060 name=collection1 org.apache.solr.core.SolrCore@67ead03 url=http://127.0.0.1:50169/collection1 node=127.0.0.1:50169_ C1060_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50169_, base_url=http://127.0.0.1:50169}
   [junit4]   2> 1831425 T4029 C1060 P50169 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=1 
   [junit4]   2> 59646 T4029 C17 P50169 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50169	GET	/schema/fields/newfield4	wt=xml	200	-	0	1	http://127.0.0.1:50169	Java/1.7.0_25	-
   [junit4]   2> 1831457 T4048 C1055 P50173 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=6 
   [junit4]   2> 59677 T4048 C22 P50173 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50173	GET	/schema/fields/newfield4	wt=xml	200	-	0	6	http://127.0.0.1:50173	Java/1.7.0_25	-
   [junit4]   2> ASYNC  NEW_CORE C1061 name=collection1 org.apache.solr.core.SolrCore@7a0eeb56 url=http://127.0.0.1:50180/collection1 node=127.0.0.1:50180_ C1061_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:50180_, base_url=http://127.0.0.1:50180}
   [junit4]   2> 1831469 T4063 C1061 P50180 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=5 
   [junit4]   2> 59689 T4063 C23 P50180 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50180	GET	/schema/fields/newfield4	wt=xml	200	-	0	5	http://127.0.0.1:50180	Java/1.7.0_25	-
   [junit4]   2> ASYNC  NEW_CORE C1062 name=collection1 org.apache.solr.core.SolrCore@66b4c55d url=http://127.0.0.1:50187/collection1 node=127.0.0.1:50187_ C1062_STATE=coll:collection1 core:collection1 props:{state=recovering, core=collection1, node_name=127.0.0.1:50187_, base_url=http://127.0.0.1:50187}
   [junit4]   2> 1831493 T4080 C1062 P50187 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield4 params={wt=xml} status=0 QTime=3 
   [junit4]   2> 59713 T4080 C24 P50187 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50187	GET	/schema/fields/newfield4	wt=xml	200	-	0	3	http://127.0.0.1:50187	Java/1.7.0_25	-
   [junit4]   2> 1831743 T3952 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831746 T4004 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831747 T4036 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831747 T4053 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831748 T3972 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831749 T4020 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831750 T3988 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831752 T4089 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831753 T4071 oass.ZkIndexSchemaReader$1.process A schema change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/configs/conf1/managed-schema, has occurred - updating schema from ZooKeeper ...
   [junit4]   2> 1831754 T4065 C1061 P50180 oass.ManagedIndexSchema.persistManagedSchemaToZooKeeper Persisted managed schema at /configs/conf1/managed-schema
   [junit4]   2> 1831755 T4065 C1061 P50180 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} status=0 QTime=234 
   [junit4]   2> 59975 T4065 C23 P50180 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50180	PUT	/schema/fields/newfield5	wt=xml	200	-	32	234	http://127.0.0.1:50180	Java/1.7.0_25	-
   [junit4]   2> 1831779 T3952 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831779 T3968 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 59999 T3968 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:08	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831793 T3952 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 14 ms
   [junit4]   2> 1831805 T4004 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831807 T3966 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60027 T3966 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831823 T4004 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 18 ms
   [junit4]   2> 1831829 T4036 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831831 T4020 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831840 T4053 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831843 T3964 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60063 T3964 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831856 T4053 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 15 ms
   [junit4]   2> 1831858 T4089 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831873 T4089 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 15 ms
   [junit4]   2> 1831875 T3965 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60095 T3965 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831880 T3988 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1831894 T3988 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 14 ms
   [junit4]   2> 1831938 T3963 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60158 T3963 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831960 T3967 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60180 T3967 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	3	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1831988 T3968 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60208 T3968 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832014 T3966 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60234 T3966 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832028 T4071 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1832046 T4071 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 17 ms
   [junit4]   2> 1832048 T3964 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60268 T3964 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832065 T3965 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=2 
   [junit4]   2> 60285 T3965 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832081 T3963 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=2 
   [junit4]   2> 60300 T3963 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832096 T3967 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60316 T3967 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832118 T3968 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=2 
   [junit4]   2> 60338 T3968 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	2	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832136 T3966 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60356 T3966 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832151 T3964 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60371 T3964 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832427 T3965 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=0 
   [junit4]   2> 60647 T3965 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832443 T3967 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=0 
   [junit4]   2> 60663 T3967 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832458 T3968 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60678 T3968 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832474 T3966 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60693 T3966 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832490 T3964 C1056 P50157 oasrs.BaseSchemaResource.handlePostExecution [collection1] webapp=null path=/schema/fields/newfield5 params={wt=xml} msg={Field 'newfield5' not found.} status=404 QTime=1 
   [junit4]   2> 60709 T3964 C18 P50157 orel.LogFilter.afterHandle 2013-07-15	05:46:09	127.0.0.1	-	127.0.0.1	50157	GET	/schema/fields/newfield5	wt=xml	404	-	0	1	http://127.0.0.1:50157	Java/1.7.0_25	-
   [junit4]   2> 1832493 T3930 oass.TestCloudManagedSchemaAddField.doTest ERROR Max retry count 20 exceeded after 734 ms.  QUERY FAILED: xpath=/response/lst[@name='responseHeader']/int[@name='status'][.='0']  request=/schema/fields/newfield5?wt=xml  response=<?xml version="1.0" encoding="UTF-8"?>
   [junit4]   2> 	<response>
   [junit4]   2> 	
   [junit4]   2> 	<lst name="responseHeader">
   [junit4]   2> 	  <int name="status">404</int>
   [junit4]   2> 	  <int name="QTime">1</int>
   [junit4]   2> 	</lst>
   [junit4]   2> 	<lst name="error">
   [junit4]   2> 	  <str name="msg">Field 'newfield5' not found.</str>
   [junit4]   2> 	  <int name="code">404</int>
   [junit4]   2> 	</lst>
   [junit4]   2> 	</response>
   [junit4]   2> 	
   [junit4]   2> 1832582 T3972 oass.ZkIndexSchemaReader.updateSchema Retrieved schema from ZooKeeper
   [junit4]   2> 1832594 T3972 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 11 ms
   [junit4]   2> 1832595 T4036 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 765 ms
   [junit4]   2> 1832599 T4020 oass.ZkIndexSchemaReader.updateSchema Finished refreshing schema in 768 ms
   [junit4]   2> 1832620 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50153
   [junit4]   2> 1832621 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1819012486
   [junit4]   2> 1833039 T4093 C1062 P50187 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:50166/collection1/ core=collection1 - recoveringAfterStartup=true
   [junit4]   2> 1833040 T4093 C1062 P50187 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:50187 START replicas=[http://127.0.0.1:50166/collection1/] nUpdates=100
   [junit4]   2> 1833043 T4093 C1062 P50187 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 1833044 T4093 C1062 P50187 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=collection1
   [junit4]   2> 1833044 T4093 C1062 P50187 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=collection1
   [junit4]   2> 1833044 T4093 C1062 P50187 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=collection1
   [junit4]   2> 1833045 T4093 C1062 P50187 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 1833045 T4093 C1062 P50187 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:50166/collection1/. core=collection1
   [junit4]   2> 1833045 T4093 C1062 P50187 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 1833082 T4012 C1059 P50166 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 1833083 T4016 C1059 P50166 oasc.SolrCore.execute [collection1] webapp= path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=1 
   [junit4]   2> 1833091 T4012 C1059 P50166 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@434e2c39 lockFactory=org.apache.lucene.store.NativeFSLockFactory@4293fe22),segFN=segments_1,generation=1}
   [junit4]   2> 1833096 T4012 C1059 P50166 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 1833098 T4012 C1059 P50166 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 1833099 T4012 C1059 P50166 oass.SolrIndexSearcher.<init> Opening Searcher@2de7ab1c realtime
   [junit4]   2> 1833100 T4012 C1059 P50166 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 1833100 T4012 C1059 P50166 oasup.LogUpdateProcessor.finish [collection1] webapp= path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 18
   [junit4]   2> 1833108 T4093 C1062 P50187 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 1833108 T4093 C1062 P50187 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 1833118 T4014 C1059 P50166 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 1833119 T4014 C1059 P50166 oasc.SolrCore.execute [collection1] webapp= path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=2 
   [junit4]   2> 1833120 T4093 C1062 P50187 oasc.RecoveryStrategy.replay No replay needed. core=collection1
   [junit4]   2> 1833120 T4093 C1062 P50187 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=collection1
   [junit4]   2> 1833121 T4093 C1062 P50187 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 1833121 T4093 C1062 P50187 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1833142 T4093 C1062 P50187 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=collection1
   [junit4]   2> 1833563 T3953 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1833566 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "node_name":"127.0.0.1:50153_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50153"}
   [junit4]   2> 1833566 T3953 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 1833575 T3953 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50187_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50187"}
   [junit4]   2> 1833588 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1833590 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1833589 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1833592 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1833591 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1833592 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1833592 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1833592 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1833592 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1834644 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1834645 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1834648 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@7cab8fc6
   [junit4]   2> 1834659 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1834659 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1834659 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1834660 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1834661 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1834662 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1834662 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data;done=false>>]
   [junit4]   2> 1834662 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data
   [junit4]   2> 1834662 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data/index;done=false>>]
   [junit4]   2> 1834663 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/control/data/index
   [junit4]   2> 1834664 T3953 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510979-127.0.0.1:50153_-n_0000000000) am no longer a leader.
   [junit4]   2> 1834670 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1834671 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1834671 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1834674 T4036 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1834676 T3972 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1834676 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834677 T3988 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1834677 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834676 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834678 T4089 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1834678 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834677 T4004 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1834678 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834686 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1834692 T4053 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1834692 T4071 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1834693 T4020 oascc.ZkStateReader$3.process Updating live nodes... (8)
   [junit4]   2> 1834694 T3972 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1834701 T3972 oasc.Overseer.start Overseer (id=90035159632510982-127.0.0.1:50157_-n_0000000001) starting
   [junit4]   2> 1834722 T4097 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1834728 T4097 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1834731 T4097 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50153_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50153"}
   [junit4]   2> 1834733 T4098 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1834740 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50157
   [junit4]   2> 1834741 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1721190892
   [junit4]   2> 1834754 T4097 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:50157_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50157"}
   [junit4]   2> 1834755 T4097 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 1834771 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834771 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834771 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834774 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834775 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834777 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834779 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1834787 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1835748 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1835749 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1835971 T3952 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1835972 T3952 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 9)
   [junit4]   2> 1835972 T3952 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1836106 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@796fdb0e
   [junit4]   2> 1836115 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=1,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1836116 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1836116 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1836117 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1836121 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1836122 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1836123 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1/index;done=false>>]
   [junit4]   2> 1836123 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1/index
   [junit4]   2> 1836124 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1 [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1;done=false>>]
   [junit4]   2> 1836124 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty1
   [junit4]   2> 1836127 T4097 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510982-127.0.0.1:50157_-n_0000000001) am no longer a leader.
   [junit4]   2> 1836195 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1836199 T4089 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1836201 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1836204 T4004 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1836206 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1836207 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1836209 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1836214 T3988 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1836214 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1836215 T4071 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1836220 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1836276 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50160
   [junit4]   2> 1836277 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1889282372
   [junit4]   2> 1836985 T4036 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 1836999 T4053 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1837000 T4020 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1837013 T4036 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 1837014 T4036 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 1837014 T4036 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1837015 T4036 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50169/collection1/
   [junit4]   2> 1837016 T4036 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:50169 START replicas=[http://127.0.0.1:50157/collection1/] nUpdates=100
   [junit4]   2> 1837017 T4036 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:50169 DONE.  We have no versions.  sync failed.
   [junit4]   2> 1837018 T4036 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 1837018 T4036 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 1837019 T4036 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50169/collection1/ shard1
   [junit4]   2> 1837019 T4036 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard1
   [junit4]   2> 1837014 T3988 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1837035 T3988 oasc.Overseer.start Overseer (id=90035159632510984-127.0.0.1:50160_-n_0000000002) starting
   [junit4]   2> 1837054 T4036 oascc.ZkStateReader$3.process Updating live nodes... (7)
   [junit4]   2> 1837054 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1837060 T4100 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1837063 T4101 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1837074 T4100 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1837078 T4100 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50157_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50157"}
   [junit4]   2> 1837089 T4100 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:50160_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50160"}
   [junit4]   2> 1837089 T4100 oasc.Overseer$ClusterStateUpdater.updateState shard=shard4 is already registered
   [junit4]   2> 1837115 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1837115 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1837116 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1837125 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1837125 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1837130 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1837133 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1837497 T3972 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1837500 T3972 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 8)
   [junit4]   2> 1837500 T3972 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1838000 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1838004 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1838007 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@7526861e
   [junit4]   2> 1838013 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=1,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1838013 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1838014 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1838014 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1838018 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1838018 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1838019 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2/index;done=false>>]
   [junit4]   2> 1838019 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2/index
   [junit4]   2> 1838020 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2 [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2;done=false>>]
   [junit4]   2> 1838020 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty2
   [junit4]   2> 1838022 T4100 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510984-127.0.0.1:50160_-n_0000000002) am no longer a leader.
   [junit4]   2> 1838027 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1838038 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1838040 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1838042 T4004 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1838042 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1838045 T4036 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1838045 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1838046 T4089 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1838046 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1838050 T4071 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1838052 T4020 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1838053 T4053 oascc.ZkStateReader$3.process Updating live nodes... (6)
   [junit4]   2> 1838063 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1838071 T4004 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1838072 T4053 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard4
   [junit4]   2> 1838090 T4053 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 1838090 T4053 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 1838091 T4053 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1838091 T4053 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50173/collection1/
   [junit4]   2> 1838091 T4053 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 1838092 T4053 oasc.SyncStrategy.syncToMe http://127.0.0.1:50173/collection1/ has no replicas
   [junit4]   2> 1838092 T4053 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50173/collection1/ shard4
   [junit4]   2> 1838092 T4053 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard4
   [junit4]   2> 1838093 T4004 oasc.Overseer.start Overseer (id=90035159632510986-127.0.0.1:50163_-n_0000000003) starting
   [junit4]   2> 1838117 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50163
   [junit4]   2> 1838118 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=502355220
   [junit4]   2> 1838136 T4102 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1838141 T4102 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1838144 T4102 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50160_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50160"}
   [junit4]   2> 1838146 T4103 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1838170 T4102 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:50163_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50163"}
   [junit4]   2> 1838172 T4102 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2> 1838207 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1838208 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1838209 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1838211 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1838213 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1838214 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1839130 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1839131 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1839135 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@2c20d3f7
   [junit4]   2> 1839146 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=1,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1839147 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1839148 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1839148 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1839152 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1839152 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1839153 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3 [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3;done=false>>]
   [junit4]   2> 1839154 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3
   [junit4]   2> 1839155 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3/index;done=false>>]
   [junit4]   2> 1839156 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty3/index
   [junit4]   2> 1839159 T4102 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510986-127.0.0.1:50163_-n_0000000003) am no longer a leader.
   [junit4]   2> 1839167 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1839168 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1839169 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1839172 T4036 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1839172 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1839173 T4089 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1839173 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1839180 T4071 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1839180 T4053 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1839181 T4020 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 1839188 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1839189 T4071 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 1839195 T4020 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1839195 T4071 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 1839196 T4071 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 1839196 T4071 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1839196 T4071 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50180/collection1/
   [junit4]   2> 1839197 T4071 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 1839197 T4071 oasc.SyncStrategy.syncToMe http://127.0.0.1:50180/collection1/ has no replicas
   [junit4]   2> 1839197 T4071 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50180/collection1/ shard2
   [junit4]   2> 1839197 T4071 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard2
   [junit4]   2> 1839207 T4020 oasc.Overseer.start Overseer (id=90035159632510988-127.0.0.1:50166_-n_0000000004) starting
   [junit4]   2> 1839234 T4105 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1839236 T4104 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1839244 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50166
   [junit4]   2> 1839244 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=800665664
   [junit4]   2> 1839248 T4104 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1839251 T4104 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50163_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50163"}
   [junit4]   2> 1839284 T4104 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:50166_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50166"}
   [junit4]   2> 1839285 T4104 oasc.Overseer$ClusterStateUpdater.updateState shard=shard3 is already registered
   [junit4]   2> 1839298 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1839303 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1839315 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1839315 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1839323 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1839340 T3988 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1839341 T3988 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 7)
   [junit4]   2> 1839341 T3988 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1840248 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1840249 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1840253 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@4300ec6e
   [junit4]   2> 1840264 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=1,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1840264 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1840265 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1840265 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1840269 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1840269 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1840270 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4 [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4;done=false>>]
   [junit4]   2> 1840270 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4
   [junit4]   2> 1840271 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4/index;done=false>>]
   [junit4]   2> 1840271 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty4/index
   [junit4]   2> 1840273 T4104 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510988-127.0.0.1:50166_-n_0000000004) am no longer a leader.
   [junit4]   2> 1840279 T4020 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1840279 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1840281 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 1840295 T4089 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard3
   [junit4]   2> 1840301 T4053 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1840301 T4036 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1840305 T4071 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1840306 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1840311 T4089 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 1840311 T4089 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 1840312 T4089 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 1840312 T4089 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50187/collection1/
   [junit4]   2> 1840312 T4089 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:50187 START replicas=[http://127.0.0.1:50166/collection1/] nUpdates=100
   [junit4]   2> 1840313 T4089 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:50187 DONE.  We have no versions.  sync failed.
   [junit4]   2> 1840314 T4089 oasc.SyncStrategy.syncReplicas Leader's attempt to sync with shard failed, moving to the next candidate
   [junit4]   2> 1840314 T4089 oasc.ShardLeaderElectionContext.runLeaderProcess We failed sync, but we have no versions - we can't sync in that case - we were active before, so become leader anyway
   [junit4]   2> 1840314 T4089 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50187/collection1/ shard3
   [junit4]   2> 1840315 T4089 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard3
   [junit4]   2> 1840314 T4036 oasc.Overseer.start Overseer (id=90035159632510990-127.0.0.1:50169_-n_0000000005) starting
   [junit4]   2> 1840349 T4089 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1840349 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1840354 T4108 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1840364 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50169
   [junit4]   2> 1840365 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=180980566
   [junit4]   2> 1840366 T4036 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 1840366 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1840369 T4107 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1840374 T4107 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1840377 T4107 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard3",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50166_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50166"}
   [junit4]   2> 1840402 T4107 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:50169_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50169"}
   [junit4]   2> 1840402 T4107 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 1840411 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1840412 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1840415 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1840415 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1840469 T4004 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1840470 T4004 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 6)
   [junit4]   2> 1840470 T4004 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1841370 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1841370 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1841373 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@67ead03
   [junit4]   2> 1841378 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1841379 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1841379 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1841380 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1841380 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1841381 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1841381 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5/index;done=false>>]
   [junit4]   2> 1841382 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5/index
   [junit4]   2> 1841382 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5 [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5;done=false>>]
   [junit4]   2> 1841382 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty5
   [junit4]   2> 1841384 T4107 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510990-127.0.0.1:50169_-n_0000000005) am no longer a leader.
   [junit4]   2> 1841394 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1841399 T4089 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 1841399 T4053 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1841400 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1841401 T4071 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 1841406 T4053 oasc.Overseer.start Overseer (id=90035159632510992-127.0.0.1:50173_-n_0000000006) starting
   [junit4]   2> 1841411 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1841419 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1841420 T4110 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1841427 T4053 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 1841429 T4109 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1841433 T4109 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1841436 T4109 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50169_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50169"}
   [junit4]   2> 1841446 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1841446 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1841449 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1841464 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50173
   [junit4]   2> 1841465 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1192201794
   [junit4]   2> 1841580 T4020 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1841581 T4020 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1842471 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1842472 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1842476 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@35c1d950
   [junit4]   2> 1842484 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1842485 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1842486 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1842486 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1842488 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1842488 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1842489 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6/index;done=false>>]
   [junit4]   2> 1842489 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6/index
   [junit4]   2> 1842490 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6 [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6;done=false>>]
   [junit4]   2> 1842490 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty6
   [junit4]   2> 1842492 T4109 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510992-127.0.0.1:50173_-n_0000000006) am no longer a leader.
   [junit4]   2> 1842501 T4089 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 1842501 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1842504 T4053 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1842509 T4071 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1842514 T4071 oasc.Overseer.start Overseer (id=90035159632510994-127.0.0.1:50180_-n_0000000007) starting
   [junit4]   2> 1842523 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1842537 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 1842538 T4112 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1842543 T4071 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 1842545 T4111 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1842549 T4111 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1842552 T4111 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:50173_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50173"}
   [junit4]   2> 1842552 T4111 oasc.Overseer$ClusterStateUpdater.updateState shard=shard4 is already registered
   [junit4]   2> 1842562 T4111 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard4",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50173_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50173"}
   [junit4]   2> 1842583 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1842584 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1842592 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50180
   [junit4]   2> 1842592 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1765733726
   [junit4]   2> 1842697 T4036 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1842698 T4036 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 1842698 T4036 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1843616 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1843617 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1843621 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@7a0eeb56
   [junit4]   2> 1843629 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1843630 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1843631 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1843631 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1843632 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1843633 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1843633 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7/index;done=false>>]
   [junit4]   2> 1843633 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7/index
   [junit4]   2> 1843634 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7 [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7;done=false>>]
   [junit4]   2> 1843634 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty7
   [junit4]   2> 1843635 T4111 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510994-127.0.0.1:50180_-n_0000000007) am no longer a leader.
   [junit4]   2> 1843642 T4071 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 1843644 T4089 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 1843650 T4089 oasc.Overseer.start Overseer (id=90035159632510996-127.0.0.1:50187_-n_0000000008) starting
   [junit4]   2> 1843663 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1843672 T4114 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 1843674 T4089 oascc.ZkStateReader$3.process Updating live nodes... (1)
   [junit4]   2> 1843674 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1843683 T4113 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 1843688 T4113 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 1843691 T4113 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:50180_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50180"}
   [junit4]   2> 1843691 T4113 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2> 1843701 T4113 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=4 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"4",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:50180_",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50180"}
   [junit4]   2> 1843710 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1843726 T3930 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 50187
   [junit4]   2> 1843727 T3930 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=614722938
   [junit4]   2> 1843806 T4053 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1843807 T4053 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1844733 T3930 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 1844733 T3930 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 1844736 T3930 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@66b4c55d
   [junit4]   2> 1844748 T3930 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=0,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=0,adds=0,deletesById=0,deletesByQuery=0,errors=0,cumulative_adds=0,cumulative_deletesById=0,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 1844749 T3930 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 1844749 T3930 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 1844750 T3930 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 1844751 T3930 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 1844751 T3930 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 1844752 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8 [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8;done=false>>]
   [junit4]   2> 1844753 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8
   [junit4]   2> 1844753 T3930 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8/index [CachedDir<<refCount=0;path=./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8/index;done=false>>]
   [junit4]   2> 1844754 T3930 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.schema.TestCloudManagedSchemaAddField-1373827508991/jetty8/index
   [junit4]   2> 1844756 T4113 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035159632510996-127.0.0.1:50187_-n_0000000008) am no longer a leader.
   [junit4]   2> 1844763 T4089 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1844763 T4089 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 1844764 T4089 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1844776 T3930 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/,null}
   [junit4]   2> 1844862 T3930 oas.SolrTestCaseJ4.tearDown ###Ending testDistribSearch
   [junit4]   2> 1844865 T3930 oasc.ZkTestServer.send4LetterWord connecting to 127.0.0.1:50149 50149
   [junit4]   2> 1844944 T4071 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 1844966 T4071 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   1> client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=TestCloudManagedSchemaAddField -Dtests.method=testDistribSearch -Dtests.seed=30BF22C5938FEB5 -Dtests.slow=true -Dtests.locale=it_IT -Dtests.timezone=Pacific/Ponape -Dtests.file.encoding=US-ASCII
   [junit4] FAILURE 73.2s | TestCloudManagedSchemaAddField.testDistribSearch <<<
   [junit4]    > Throwable #1: java.lang.AssertionError: Max retry count 20 exceeded after 734 ms.  QUERY FAILED: xpath=/response/lst[@name='responseHeader']/int[@name='status'][.='0']  request=/schema/fields/newfield5?wt=xml  response=<?xml version="1.0" encoding="UTF-8"?>
   [junit4]    > <response>
   [junit4]    > <lst name="responseHeader">
   [junit4]    >   <int name="status">404</int>
   [junit4]    >   <int name="QTime">1</int>
   [junit4]    > </lst>
   [junit4]    > <lst name="error">
   [junit4]    >   <str name="msg">Field 'newfield5' not found.</str>
   [junit4]    >   <int name="code">404</int>
   [junit4]    > </lst>
   [junit4]    > </response>
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([30BF22C5938FEB5:82ED7C342E679E89]:0)
   [junit4]    > 	at org.apache.solr.schema.TestCloudManagedSchemaAddField.doTest(TestCloudManagedSchemaAddField.java:119)
   [junit4]    > 	at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:835)
   [junit4]    > 	at java.lang.Thread.run(Thread.java:724)
   [junit4]   2> 1844981 T3930 oas.SolrTestCaseJ4.deleteCore ###deleteCore
   [junit4]   2> 73208 T3929 ccr.ThreadLeakControl.checkThreadLeaks WARNING Will linger awaiting termination of 1 leaked thread(s).
   [junit4]   2> NOTE: test params are: codec=Lucene42: {}, docValues:{}, sim=DefaultSimilarity, locale=it_IT, timezone=Pacific/Ponape
   [junit4]   2> NOTE: Mac OS X 10.8.4 x86_64/Oracle Corporation 1.7.0_25 (64-bit)/cpus=2,threads=1,free=108930120,total=357433344
   [junit4]   2> NOTE: All tests run in this JVM: [BadComponentTest, DistributedQueryElevationComponentTest, RAMDirectoryFactoryTest, TestReloadAndDeleteDocs, IndexSchemaTest, SpellCheckComponentTest, PeerSyncTest, ZkControllerTest, TestStressVersions, DistributedTermsComponentTest, ChaosMonkeyNothingIsSafeTest, FieldMutatingUpdateProcessorTest, LegacyHTMLStripCharFilterTest, TestSolrXMLSerializer, SOLR749Test, TestSolrQueryParser, TestFieldTypeCollectionResource, PreAnalyzedUpdateProcessorTest, TestSolrJ, TestCloudManagedSchema, TestQuerySenderNoQuery, HdfsBasicDistributedZk2Test, ResponseLogComponentTest, TestQuerySenderListener, TestWriterPerf, CoreAdminCreateDiscoverTest, SearchHandlerTest, TestPostingsSolrHighlighter, TestClassNameShortening, CollectionsAPIDistributedZkTest, SoftAutoCommitTest, DistanceFunctionTest, TestConfig, DocumentAnalysisRequestHandlerTest, ZkNodePropsTest, TestCSVLoader, SuggesterTSTTest, TestGroupingSearch, BasicDistributedZkTest, TestSweetSpotSimilarityFactory, TestManagedSchema, OpenCloseCoreStressTest, AddSchemaFieldsUpdateProcessorFactoryTest, SuggesterTest, TestSolrCoreProperties, TestAddFieldRealTimeGet, TestLazyCores, SpellPossibilityIteratorTest, TestSort, TestHashPartitioner, SolrCoreTest, FullSolrCloudDistribCmdsTest, TestTrie, DebugComponentTest, TestPartialUpdateDeduplication, TestReplicationHandler, HdfsSyncSliceTest, ClusterStateTest, SchemaVersionSpecificBehaviorTest, CSVRequestHandlerTest, ConvertedLegacyTest, AlternateDirectoryTest, PolyFieldTest, PluginInfoTest, TestDynamicFieldResource, TestFieldTypeResource, TestValueSourceCache, QueryResultKeyTest, BasicFunctionalityTest, AnalysisAfterCoreReloadTest, TestSchemaResource, TestSuggestSpellingConverter, ChaosMonkeySafeLeaderTest, TestJmxMonitoredMap, HardAutoCommitTest, TestSolrXmlPersistence, TestSchemaSimilarityResource, TestDefaultSimilarityFactory, TestInfoStreamLogging, TestPhraseSuggestions, AliasIntegrationTest, TestRandomDVFaceting, TestFuzzyAnalyzedSuggestions, TestQueryUtils, TestSchemaNameResource, UpdateRequestProcessorFactoryTest, TestPHPSerializedResponseWriter, ScriptEngineTest, TestSystemIdResolver, TestQueryTypes, MultiTermTest, TestFastWriter, SignatureUpdateProcessorFactoryTest, TestSolrQueryParserResource, PrimUtilsTest, HdfsBasicDistributedZkTest, TestIBSimilarityFactory, ParsingFieldUpdateProcessorsTest, TestPerFieldSimilarity, URLClassifyProcessorTest, CoreContainerCoreInitFailuresTest, TestZkChroot, ClusterStateUpdateTest, TestManagedSchemaFieldResource, SimplePostToolTest, TestCodecSupport, JsonLoaderTest, TestJmxIntegration, DocValuesTest, RequestHandlersTest, DefaultValueUpdateProcessorTest, DocValuesMultiTest, QueryParsingTest, ExternalFileFieldSortTest, TestCharFilters, TestSchemaVersionResource, TestXIncludeConfig, StatsComponentTest, CacheHeaderTest, PathHierarchyTokenizerFactoryTest, TestPropInjectDefaults, IndexBasedSpellCheckerTest, SystemInfoHandlerTest, TestCloudManagedSchemaAddField]
   [junit4] Completed in 74.34s, 1 test, 1 failure <<< FAILURES!

[...truncated 503 lines...]
   [junit4] Suite: org.apache.solr.cloud.ShardSplitTest
   [junit4]   2> 3928983 T7990 oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system property: /qb/
   [junit4]   2> 3928989 T7990 oas.SolrTestCaseJ4.setUp ###Starting testDistribSearch
   [junit4]   2> Creating dataDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./solrtest-ShardSplitTest-1373829666186
   [junit4]   2> 3928991 T7990 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   2> 3928993 T7991 oasc.ZkTestServer$ZKServerMain.runFromConfig Starting server
   [junit4]   2> 3929093 T7990 oasc.ZkTestServer.run start zk server on port:51248
   [junit4]   2> 3929095 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3929103 T7997 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@5559d11c name:ZooKeeperConnection Watcher:127.0.0.1:51248 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3929104 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3929105 T7990 oascc.SolrZkClient.makePath makePath: /solr
   [junit4]   2> 3929117 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3929123 T7999 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@53eac41b name:ZooKeeperConnection Watcher:127.0.0.1:51248/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3929124 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3929125 T7990 oascc.SolrZkClient.makePath makePath: /collections/collection1
   [junit4]   2> 3929132 T7990 oascc.SolrZkClient.makePath makePath: /collections/collection1/shards
   [junit4]   2> 3929139 T7990 oascc.SolrZkClient.makePath makePath: /collections/control_collection
   [junit4]   2> 3929144 T7990 oascc.SolrZkClient.makePath makePath: /collections/control_collection/shards
   [junit4]   2> 3929151 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml
   [junit4]   2> 3929152 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/solrconfig.xml
   [junit4]   2> 3929172 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/schema15.xml to /configs/conf1/schema.xml
   [junit4]   2> 3929173 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/schema.xml
   [junit4]   2> 3929183 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
   [junit4]   2> 3929185 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/stopwords.txt
   [junit4]   2> 3929197 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
   [junit4]   2> 3929199 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/protwords.txt
   [junit4]   2> 3929207 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
   [junit4]   2> 3929208 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/currency.xml
   [junit4]   2> 3929217 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 3929218 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/open-exchange-rates.json
   [junit4]   2> 3929224 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 3929225 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 3929233 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
   [junit4]   2> 3929234 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/old_synonyms.txt
   [junit4]   2> 3929241 T7990 oasc.AbstractZkTestCase.putConfig put /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
   [junit4]   2> 3929242 T7990 oascc.SolrZkClient.makePath makePath: /configs/conf1/synonyms.txt
   [junit4]   2> 3929821 T7990 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 3929836 T7990 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:51251
   [junit4]   2> 3929838 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 3929838 T7990 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 3929842 T7990 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446
   [junit4]   2> 3929843 T7990 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/'
   [junit4]   2> 3929895 T7990 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/solr.xml
   [junit4]   2> 3930011 T7990 oasc.CoreContainer.<init> New CoreContainer 1171924669
   [junit4]   2> 3930012 T7990 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/]
   [junit4]   2> 3930014 T7990 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 3930014 T7990 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 3930015 T7990 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 3930015 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 3930016 T7990 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 3930016 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 3930017 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 3930017 T7990 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 3930018 T7990 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 3930018 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 3930046 T7990 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 3930047 T7990 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:51248/solr
   [junit4]   2> 3930047 T7990 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 3930050 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3930055 T8010 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@dd4ab7a name:ZooKeeperConnection Watcher:127.0.0.1:51248 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3930056 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3930060 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 3930089 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3930095 T8012 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@18baec10 name:ZooKeeperConnection Watcher:127.0.0.1:51248/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3930095 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3930104 T7990 oascc.SolrZkClient.makePath makePath: /overseer/queue
   [junit4]   2> 3930113 T7990 oascc.SolrZkClient.makePath makePath: /overseer/collection-queue-work
   [junit4]   2> 3930120 T7990 oascc.SolrZkClient.makePath makePath: /live_nodes
   [junit4]   2> 3930126 T7990 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:51251_qb
   [junit4]   2> 3930130 T7990 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:51251_qb
   [junit4]   2> 3930137 T7990 oascc.SolrZkClient.makePath makePath: /overseer_elect/election
   [junit4]   2> 3930147 T7990 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 3930152 T7990 oasc.Overseer.start Overseer (id=90035301003689987-127.0.0.1:51251_qb-n_0000000000) starting
   [junit4]   2> 3930159 T7990 oascc.SolrZkClient.makePath makePath: /overseer/queue-work
   [junit4]   2> 3930170 T8014 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 3930175 T7990 oascc.SolrZkClient.makePath makePath: /clusterstate.json
   [junit4]   2> 3930181 T7990 oascc.SolrZkClient.makePath makePath: /aliases.json
   [junit4]   2> 3930188 T7990 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 3930199 T8013 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 3930203 T8015 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 3930203 T8015 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3930205 T8015 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 3931708 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3931710 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=1 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"1",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3931710 T8013 oasc.Overseer$ClusterStateUpdater.createCollection Create collection control_collection with numShards 1
   [junit4]   2> 3931710 T8013 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard1
   [junit4]   2> 3931717 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 3932208 T8015 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 3932209 T8015 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1
   [junit4]   2> 3932210 T8015 oasc.ZkController.createCollectionZkNode Check for collection zkNode:control_collection
   [junit4]   2> 3932211 T8015 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3932212 T8015 oasc.ZkController.readConfigName Load collection config from:/collections/control_collection
   [junit4]   2> 3932215 T8015 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1/'
   [junit4]   2> 3932217 T8015 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1/lib/classes/' to classloader
   [junit4]   2> 3932218 T8015 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1/lib/README' to classloader
   [junit4]   2> 3932305 T8015 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 3932393 T8015 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3932396 T8015 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3932410 T8015 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 3933208 T8015 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 3933219 T8015 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 3933222 T8015 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3933229 T8015 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3933289 T8015 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3933290 T8015 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1/, dataDir=./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data/
   [junit4]   2> 3933290 T8015 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3933292 T8015 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data
   [junit4]   2> 3933292 T8015 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data/index/
   [junit4]   2> 3933292 T8015 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data/index' doesn't exist. Creating new index...
   [junit4]   2> 3933293 T8015 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data/index
   [junit4]   2> 3933298 T8015 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@65e17bbd lockFactory=org.apache.lucene.store.NativeFSLockFactory@7c5b2df5),segFN=segments_1,generation=1}
   [junit4]   2> 3933299 T8015 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3933304 T8015 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3933305 T8015 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3933307 T8015 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3933307 T8015 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3933308 T8015 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3933308 T8015 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3933308 T8015 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3933309 T8015 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3933309 T8015 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3933323 T8015 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3933331 T8015 oass.SolrIndexSearcher.<init> Opening Searcher@4bead742 main
   [junit4]   2> 3933332 T8015 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3933333 T8015 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3933338 T8016 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@4bead742 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3933339 T8015 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 3933339 T8015 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:51251/qb collection:control_collection shard:shard1
   [junit4]   2> 3933340 T8015 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3933365 T8015 oascc.SolrZkClient.makePath makePath: /collections/control_collection/leader_elect/shard1/election
   [junit4]   2> 3933380 T8015 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 3933386 T8015 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 3933387 T8015 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 3933387 T8015 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:51251/qb/collection1/
   [junit4]   2> 3933387 T8015 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 3933388 T8015 oasc.SyncStrategy.syncToMe http://127.0.0.1:51251/qb/collection1/ has no replicas
   [junit4]   2> 3933388 T8015 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:51251/qb/collection1/ shard1
   [junit4]   2> 3933389 T8015 oascc.SolrZkClient.makePath makePath: /collections/control_collection/leaders/shard1
   [junit4]   2> 3934732 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3934752 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 3934788 T8015 oasc.ZkController.register We are http://127.0.0.1:51251/qb/collection1/ and leader is http://127.0.0.1:51251/qb/collection1/
   [junit4]   2> 3934789 T8015 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 3934789 T8015 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 3934789 T8015 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3934793 T8015 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3934796 T7990 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 3934797 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 3934798 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3934814 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:
   [junit4]   2> 3934820 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3934827 T8019 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@67f42fd9 name:ZooKeeperConnection Watcher:127.0.0.1:51248/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3934828 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3934831 T7990 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 3934837 T7990 oasc.ChaosMonkey.monkeyLog monkey: init - expire sessions:true cause connection loss:true
   [junit4]   2> 3935329 T7990 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 3935335 T7990 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:51255
   [junit4]   2> 3935337 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 3935338 T7990 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 3935338 T7990 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033
   [junit4]   2> 3935339 T7990 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/'
   [junit4]   2> 3935405 T7990 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/solr.xml
   [junit4]   2> 3935513 T7990 oasc.CoreContainer.<init> New CoreContainer 939315043
   [junit4]   2> 3935515 T7990 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/]
   [junit4]   2> 3935522 T7990 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 3935523 T7990 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 3935524 T7990 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 3935524 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 3935525 T7990 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 3935525 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 3935526 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 3935526 T7990 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 3935527 T7990 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 3935527 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 3935546 T7990 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 3935547 T7990 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:51248/solr
   [junit4]   2> 3935548 T7990 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 3935551 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3935554 T8030 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@5ab54180 name:ZooKeeperConnection Watcher:127.0.0.1:51248 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3935555 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3935558 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 3935576 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3935583 T8032 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@685371ad name:ZooKeeperConnection Watcher:127.0.0.1:51248/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3935584 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3935600 T7990 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 3936271 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3936275 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=1 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"1",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3936287 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 3936288 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 3936295 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 3936612 T7990 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:51255_qb
   [junit4]   2> 3936683 T7990 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:51255_qb
   [junit4]   2> 3936691 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 3936692 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 3936701 T8019 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 3936702 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3936710 T8012 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 3936715 T8032 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 3936749 T8033 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 3936749 T8033 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3936757 T8033 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 3937805 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3937807 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3937808 T8013 oasc.Overseer$ClusterStateUpdater.createCollection Create collection collection1 with numShards 2
   [junit4]   2> 3937808 T8013 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard1
   [junit4]   2> 3937819 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3937823 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3937824 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3938760 T8033 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 3938760 T8033 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1
   [junit4]   2> 3938760 T8033 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 3938762 T8033 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3938762 T8033 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 3938764 T8033 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1/'
   [junit4]   2> 3938766 T8033 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1/lib/classes/' to classloader
   [junit4]   2> 3938766 T8033 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1/lib/README' to classloader
   [junit4]   2> 3938852 T8033 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 3938937 T8033 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3938940 T8033 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3938964 T8033 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 3939793 T8033 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 3939799 T8033 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 3939802 T8033 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3939809 T8033 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3939869 T8033 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3939869 T8033 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1/, dataDir=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1/
   [junit4]   2> 3939870 T8033 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3939871 T8033 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1
   [junit4]   2> 3939872 T8033 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1/index/
   [junit4]   2> 3939872 T8033 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1/index' doesn't exist. Creating new index...
   [junit4]   2> 3939873 T8033 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1/index
   [junit4]   2> 3939877 T8033 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_1,generation=1}
   [junit4]   2> 3939878 T8033 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3939894 T8033 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3939895 T8033 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3939896 T8033 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3939897 T8033 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3939898 T8033 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3939899 T8033 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3939899 T8033 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3939900 T8033 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3939901 T8033 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3939936 T8033 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3939979 T8033 oass.SolrIndexSearcher.<init> Opening Searcher@284d7bd7 main
   [junit4]   2> 3939982 T8033 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3939983 T8033 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3939992 T8034 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@284d7bd7 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3939995 T8033 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 3939996 T8033 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:51255/qb collection:collection1 shard:shard1
   [junit4]   2> 3939996 T8033 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3940020 T8033 oascc.SolrZkClient.makePath makePath: /collections/collection1/leader_elect/shard1/election
   [junit4]   2> 3940066 T8033 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 3940100 T8033 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 3940100 T8033 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 3940101 T8033 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:51255/qb/collection1/
   [junit4]   2> 3940101 T8033 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 3940102 T8033 oasc.SyncStrategy.syncToMe http://127.0.0.1:51255/qb/collection1/ has no replicas
   [junit4]   2> 3940102 T8033 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:51255/qb/collection1/ shard1
   [junit4]   2> 3940102 T8033 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard1
   [junit4]   2> 3940852 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3940899 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3940899 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3940899 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3940914 T8033 oasc.ZkController.register We are http://127.0.0.1:51255/qb/collection1/ and leader is http://127.0.0.1:51255/qb/collection1/
   [junit4]   2> 3940914 T8033 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 3940915 T8033 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 3940915 T8033 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3940919 T8033 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3940925 T7990 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 3940927 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 3940927 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3941706 T7990 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 3941729 T7990 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:51258
   [junit4]   2> 3941730 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 3941731 T7990 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 3941731 T7990 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140
   [junit4]   2> 3941732 T7990 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/'
   [junit4]   2> 3941799 T7990 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/solr.xml
   [junit4]   2> 3941991 T7990 oasc.CoreContainer.<init> New CoreContainer 848682637
   [junit4]   2> 3941993 T7990 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/]
   [junit4]   2> 3941996 T7990 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 3941996 T7990 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 3941997 T7990 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 3941998 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 3941998 T7990 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 3941999 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 3941999 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 3942000 T7990 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 3942000 T7990 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 3942001 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 3942023 T7990 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 3942024 T7990 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:51248/solr
   [junit4]   2> 3942024 T7990 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 3942042 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3942054 T8046 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@2f945f02 name:ZooKeeperConnection Watcher:127.0.0.1:51248 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3942056 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3942063 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 3942087 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3942094 T8048 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@18dd83a1 name:ZooKeeperConnection Watcher:127.0.0.1:51248/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3942095 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3942132 T7990 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 3942447 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3942450 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3942472 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3942472 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3942472 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3942473 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3943151 T7990 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:51258_qb
   [junit4]   2> 3943161 T7990 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:51258_qb
   [junit4]   2> 3943172 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3943173 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 3943176 T8048 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 3943176 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3943176 T8019 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 3943177 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3943181 T8032 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 3943184 T8012 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 3943200 T8049 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 3943200 T8049 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3943205 T8049 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 3943999 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3944004 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 3944004 T8013 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=2
   [junit4]   2> 3944005 T8013 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard2
   [junit4]   2> 3944019 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3944019 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3944019 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3944020 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3944207 T8049 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 3944208 T8049 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1
   [junit4]   2> 3944208 T8049 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 3944210 T8049 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3944211 T8049 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 3944213 T8049 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1/'
   [junit4]   2> 3944215 T8049 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1/lib/classes/' to classloader
   [junit4]   2> 3944215 T8049 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1/lib/README' to classloader
   [junit4]   2> 3944319 T8049 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 3944433 T8049 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3944437 T8049 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3944453 T8049 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 3945580 T8049 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 3945588 T8049 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 3945594 T8049 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3945617 T8049 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3945770 T8049 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3945771 T8049 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1/, dataDir=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2/
   [junit4]   2> 3945772 T8049 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3945774 T8049 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2
   [junit4]   2> 3945779 T8049 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2/index/
   [junit4]   2> 3945779 T8049 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2/index' doesn't exist. Creating new index...
   [junit4]   2> 3945791 T8049 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2/index
   [junit4]   2> 3945812 T8049 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_1,generation=1}
   [junit4]   2> 3945813 T8049 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3945820 T8049 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3945821 T8049 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3945822 T8049 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3945823 T8049 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3945824 T8049 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3945825 T8049 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3945825 T8049 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3945826 T8049 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3945827 T8049 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3945847 T8049 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3945865 T8049 oass.SolrIndexSearcher.<init> Opening Searcher@2ac47070 main
   [junit4]   2> 3945871 T8049 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3945872 T8049 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3945878 T8050 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@2ac47070 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3945883 T8049 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 3945884 T8049 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:51258/qb collection:collection1 shard:shard2
   [junit4]   2> 3945884 T8049 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3945908 T8049 oascc.SolrZkClient.makePath makePath: /collections/collection1/leader_elect/shard2/election
   [junit4]   2> 3945945 T8049 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 3945962 T8049 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 3945962 T8049 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 3945963 T8049 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:51258/qb/collection1/
   [junit4]   2> 3945963 T8049 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 3945963 T8049 oasc.SyncStrategy.syncToMe http://127.0.0.1:51258/qb/collection1/ has no replicas
   [junit4]   2> 3945964 T8049 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:51258/qb/collection1/ shard2
   [junit4]   2> 3945966 T8049 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard2
   [junit4]   2> 3947044 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3947087 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3947087 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3947088 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3947087 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3947134 T8049 oasc.ZkController.register We are http://127.0.0.1:51258/qb/collection1/ and leader is http://127.0.0.1:51258/qb/collection1/
   [junit4]   2> 3947134 T8049 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 3947134 T8049 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 3947134 T8049 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3947137 T8049 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3947140 T7990 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 3947140 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 3947141 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3947916 T7990 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 3947934 T7990 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:51261
   [junit4]   2> 3947938 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 3947939 T7990 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 3947940 T7990 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357
   [junit4]   2> 3947941 T7990 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357/'
   [junit4]   2> 3948005 T7990 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357/solr.xml
   [junit4]   2> 3948160 T7990 oasc.CoreContainer.<init> New CoreContainer 1975274070
   [junit4]   2> 3948161 T7990 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357/]
   [junit4]   2> 3948164 T7990 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 3948164 T7990 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 3948165 T7990 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 3948165 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 3948166 T7990 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 3948166 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 3948167 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 3948167 T7990 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 3948168 T7990 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 3948168 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 3948196 T7990 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 3948197 T7990 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:51248/solr
   [junit4]   2> 3948198 T7990 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 3948200 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3948208 T8062 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@afcf4d6 name:ZooKeeperConnection Watcher:127.0.0.1:51248 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3948210 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3948229 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 3948246 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3948253 T8064 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@1f463c75 name:ZooKeeperConnection Watcher:127.0.0.1:51248/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3948257 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3948299 T7990 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 3948614 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3948622 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 3948636 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3948637 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3948637 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3948637 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3948637 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3949331 T7990 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:51261_qb
   [junit4]   2> 3949336 T7990 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:51261_qb
   [junit4]   2> 3949344 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3949345 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3949349 T8019 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3949350 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3949350 T8048 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3949351 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3949351 T8064 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3949351 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3949357 T8012 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3949357 T8032 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3949377 T8065 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 3949378 T8065 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3949381 T8065 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 3950185 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3950188 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51261_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51261/qb"}
   [junit4]   2> 3950188 T8013 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=2
   [junit4]   2> 3950189 T8013 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard1
   [junit4]   2> 3950200 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3950204 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3950205 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3950204 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3950204 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3950383 T8065 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 3950384 T8065 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357/collection1
   [junit4]   2> 3950384 T8065 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 3950385 T8065 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3950386 T8065 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 3950388 T8065 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357/collection1/'
   [junit4]   2> 3950389 T8065 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357/collection1/lib/classes/' to classloader
   [junit4]   2> 3950390 T8065 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357/collection1/lib/README' to classloader
   [junit4]   2> 3950463 T8065 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 3950546 T8065 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3950549 T8065 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3950560 T8065 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 3951329 T8065 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 3951335 T8065 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 3951337 T8065 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3951346 T8065 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3951412 T8065 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3951413 T8065 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty3-1373829684357/collection1/, dataDir=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3/
   [junit4]   2> 3951413 T8065 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3951414 T8065 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3
   [junit4]   2> 3951414 T8065 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3/index/
   [junit4]   2> 3951415 T8065 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3/index' doesn't exist. Creating new index...
   [junit4]   2> 3951415 T8065 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3/index
   [junit4]   2> 3951419 T8065 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@4cf6277a lockFactory=org.apache.lucene.store.NativeFSLockFactory@518a7bdf),segFN=segments_1,generation=1}
   [junit4]   2> 3951420 T8065 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3951428 T8065 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3951430 T8065 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3951430 T8065 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3951431 T8065 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3951432 T8065 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3951433 T8065 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3951433 T8065 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3951435 T8065 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3951435 T8065 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3951455 T8065 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3951465 T8065 oass.SolrIndexSearcher.<init> Opening Searcher@2cd7adaa main
   [junit4]   2> 3951467 T8065 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3951467 T8065 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3951474 T8066 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@2cd7adaa main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3951478 T8065 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 3951479 T8065 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:51261/qb collection:collection1 shard:shard1
   [junit4]   2> 3951479 T8065 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3951502 T8065 oasc.ZkController.register We are http://127.0.0.1:51261/qb/collection1/ and leader is http://127.0.0.1:51255/qb/collection1/
   [junit4]   2> 3951503 T8065 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:51261/qb
   [junit4]   2> 3951503 T8065 oasc.ZkController.checkRecovery Core needs to recover:collection1
   [junit4]   2> 3951503 T8065 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> ASYNC  NEW_CORE C1813 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1813_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3951504 T8067 C1813 P51261 oasc.RecoveryStrategy.run Starting recovery process.  core=collection1 recoveringAfterStartup=true
   [junit4]   2> 3951506 T8067 C1813 P51261 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 3951506 T8067 C1813 P51261 oasc.ZkController.publish publishing core=collection1 state=recovering
   [junit4]   2> 3951507 T8067 C1813 P51261 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3951510 T8067 C1813 P51261 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3951517 T8065 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3951521 T7990 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 3951521 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 3951522 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3951554 T8026 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node3, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 3951714 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3951721 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51261_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51261/qb"}
   [junit4]   2> 3951748 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3951749 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3951748 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3951749 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3951749 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3952347 T7990 oejs.Server.doStart jetty-8.1.10.v20130312
   [junit4]   2> 3952375 T7990 oejs.AbstractConnector.doStart Started SelectChannelConnector@127.0.0.1:51265
   [junit4]   2> 3952381 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init()
   [junit4]   2> 3952382 T7990 oasc.SolrResourceLoader.locateSolrHome JNDI not configured for solr (NoInitialContextEx)
   [junit4]   2> 3952382 T7990 oasc.SolrResourceLoader.locateSolrHome using system property solr.solr.home: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743
   [junit4]   2> 3952383 T7990 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743/'
   [junit4]   2> 3952445 T7990 oasc.ConfigSolr.fromFile Loading container configuration from /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743/solr.xml
   [junit4]   2> 3952559 T8026 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node3, state: recovering, checkLive: true, onlyIfLeader: true for: 1 seconds.
   [junit4]   2> 3952560 T8026 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node3&state=recovering&nodeName=127.0.0.1:51261_qb&action=PREPRECOVERY&checkLive=true&core=collection1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=1006 
   [junit4]   2> 3952635 T7990 oasc.CoreContainer.<init> New CoreContainer 1583629964
   [junit4]   2> 3952638 T7990 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743/]
   [junit4]   2> 3952647 T7990 oashc.HttpShardHandlerFactory.getParameter Setting socketTimeout to: 120000
   [junit4]   2> 3952648 T7990 oashc.HttpShardHandlerFactory.getParameter Setting urlScheme to: http://
   [junit4]   2> 3952649 T7990 oashc.HttpShardHandlerFactory.getParameter Setting connTimeout to: 15000
   [junit4]   2> 3952650 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxConnectionsPerHost to: 20
   [junit4]   2> 3952651 T7990 oashc.HttpShardHandlerFactory.getParameter Setting corePoolSize to: 0
   [junit4]   2> 3952652 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maximumPoolSize to: 2147483647
   [junit4]   2> 3952653 T7990 oashc.HttpShardHandlerFactory.getParameter Setting maxThreadIdleTime to: 5
   [junit4]   2> 3952654 T7990 oashc.HttpShardHandlerFactory.getParameter Setting sizeOfQueue to: -1
   [junit4]   2> 3952655 T7990 oashc.HttpShardHandlerFactory.getParameter Setting fairnessPolicy to: false
   [junit4]   2> 3952656 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnectionsPerHost=20&maxConnections=10000&socketTimeout=120000&connTimeout=15000&retry=false
   [junit4]   2> 3952691 T7990 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 3952692 T7990 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:51248/solr
   [junit4]   2> 3952693 T7990 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 3952695 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3952803 T8079 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@17cc8469 name:ZooKeeperConnection Watcher:127.0.0.1:51248 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3952804 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3952845 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=500&maxConnectionsPerHost=16&socketTimeout=120000&connTimeout=15000
   [junit4]   2> 3952877 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3952885 T8081 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@1333748e name:ZooKeeperConnection Watcher:127.0.0.1:51248/solr got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3952886 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3952909 T7990 oascc.ZkStateReader.createClusterStateWatchersAndUpdate Updating cluster state from ZooKeeper... 
   [junit4]   2> 3953932 T7990 oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:51265_qb
   [junit4]   2> 3953935 T7990 oascc.SolrZkClient.makePath makePath: /live_nodes/127.0.0.1:51265_qb
   [junit4]   2> 3953955 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3953956 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3953958 T8048 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 3953959 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3953959 T8081 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 3953960 T8064 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 3953960 T8019 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 3953962 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3953961 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3953966 T8012 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 3953967 T8032 oascc.ZkStateReader$3.process Updating live nodes... (5)
   [junit4]   2> 3953983 T8082 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 3953983 T8082 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3953986 T8082 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> ASYNC  NEW_CORE C1814 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1814_STATE=coll:collection1 core:collection1 props:{state=recovering, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3954562 T8067 C1814 P51261 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:51255/qb/collection1/ core=collection1 - recoveringAfterStartup=true
   [junit4]   2> 3954563 T8067 C1814 P51261 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:51261/qb START replicas=[http://127.0.0.1:51255/qb/collection1/] nUpdates=100
   [junit4]   2> 3954564 T8067 C1814 P51261 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 3954564 T8067 C1814 P51261 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=collection1
   [junit4]   2> 3954565 T8067 C1814 P51261 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=collection1
   [junit4]   2> 3954565 T8067 C1814 P51261 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=collection1
   [junit4]   2> 3954565 T8067 C1814 P51261 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 3954566 T8067 C1814 P51261 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:51255/qb/collection1/. core=collection1
   [junit4]   2> 3954570 T8067 C1814 P51261 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> ASYNC  NEW_CORE C1815 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1815_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3954603 T8023 C1815 P51255 oasc.SolrCore.execute [collection1] webapp=/qb path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=21 
   [junit4]   2> 3954608 T8024 C1815 P51255 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3954612 T8024 C1815 P51255 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_1,generation=1}
   [junit4]   2> 3954614 T8024 C1815 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3954615 T8024 C1815 P51255 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 3954617 T8024 C1815 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@71a72d61 realtime
   [junit4]   2> 3954617 T8024 C1815 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3954617 T8024 C1815 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 9
   [junit4]   2> 3954618 T8067 C1814 P51261 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 3954619 T8067 C1814 P51261 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 3954624 T8027 C1815 P51255 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 3954624 T8027 C1815 P51255 oasc.SolrCore.execute [collection1] webapp=/qb path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=2 
   [junit4]   2> 3954627 T8067 C1814 P51261 oasc.RecoveryStrategy.replay No replay needed. core=collection1
   [junit4]   2> 3954627 T8067 C1814 P51261 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=collection1
   [junit4]   2> 3954628 T8067 C1814 P51261 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 3954628 T8067 C1814 P51261 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3954633 T8067 C1814 P51261 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=collection1
   [junit4]   2> 3954787 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3954790 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51265_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51265/qb"}
   [junit4]   2> 3954791 T8013 oasc.Overseer$ClusterStateUpdater.updateState Collection already exists with numShards=2
   [junit4]   2> 3954791 T8013 oasc.Overseer$ClusterStateUpdater.updateState Assigning new node to shard shard=shard2
   [junit4]   2> 3954804 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51261_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51261/qb"}
   [junit4]   2> 3954815 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3954815 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3954816 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3954815 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3954816 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3954815 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3954990 T8082 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 3954990 T8082 oasc.CoreContainer.create Creating SolrCore 'collection1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743/collection1
   [junit4]   2> 3954991 T8082 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 3954992 T8082 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3954992 T8082 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 3954994 T8082 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743/collection1/'
   [junit4]   2> 3954996 T8082 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743/collection1/lib/classes/' to classloader
   [junit4]   2> 3954997 T8082 oasc.SolrResourceLoader.replaceClassLoader Adding 'file:/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743/collection1/lib/README' to classloader
   [junit4]   2> 3955078 T8082 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 3955169 T8082 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3955172 T8082 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3955189 T8082 oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 3955994 T8082 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 3955998 T8082 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 3956001 T8082 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3956007 T8082 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3956069 T8082 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3956070 T8082 oasc.SolrCore.<init> [collection1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty4-1373829688743/collection1/, dataDir=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4/
   [junit4]   2> 3956071 T8082 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3956072 T8082 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4
   [junit4]   2> 3956072 T8082 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4/index/
   [junit4]   2> 3956073 T8082 oasc.SolrCore.initIndex WARN [collection1] Solr index directory './org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4/index' doesn't exist. Creating new index...
   [junit4]   2> 3956074 T8082 oasc.CachingDirectoryFactory.get return new directory for ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4/index
   [junit4]   2> 3956079 T8082 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191fa213 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1b45ecd0),segFN=segments_1,generation=1}
   [junit4]   2> 3956079 T8082 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3956088 T8082 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3956089 T8082 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3956091 T8082 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3956092 T8082 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3956093 T8082 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3956093 T8082 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3956094 T8082 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3956095 T8082 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3956095 T8082 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3956116 T8082 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3956124 T8082 oass.SolrIndexSearcher.<init> Opening Searcher@43d080d9 main
   [junit4]   2> 3956125 T8082 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3956126 T8082 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3956134 T8084 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@43d080d9 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3956137 T8082 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 3956137 T8082 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:51265/qb collection:collection1 shard:shard2
   [junit4]   2> 3956138 T8082 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3956170 T8082 oasc.ZkController.register We are http://127.0.0.1:51265/qb/collection1/ and leader is http://127.0.0.1:51258/qb/collection1/
   [junit4]   2> 3956171 T8082 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:51265/qb
   [junit4]   2> 3956171 T8082 oasc.ZkController.checkRecovery Core needs to recover:collection1
   [junit4]   2> 3956172 T8082 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> ASYNC  NEW_CORE C1816 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1816_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3956173 T8085 C1816 P51265 oasc.RecoveryStrategy.run Starting recovery process.  core=collection1 recoveringAfterStartup=true
   [junit4]   2> 3956174 T8082 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3956174 T8085 C1816 P51265 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 3956174 T8085 C1816 P51265 oasc.ZkController.publish publishing core=collection1 state=recovering
   [junit4]   2> 3956175 T8085 C1816 P51265 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3956178 T8085 C1816 P51265 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3956178 T7990 oass.SolrDispatchFilter.init user.dir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0
   [junit4]   2> 3956179 T7990 oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 3956180 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3956196 T7990 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3956199 T7990 oasc.AbstractFullDistribZkTestBase.waitForThingsToLevelOut Wait for recoveries to finish - wait 15 for each attempt
   [junit4]   2> 3956199 T7990 oasc.AbstractDistribZkTestBase.waitForRecoveriesToFinish Wait for recoveries to finish - collection: collection1 failOnTimeout:true timeout (sec):15
   [junit4]   2> 3956201 T8042 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node4, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 3956202 T7990 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3956339 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3956344 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51265_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51265/qb"}
   [junit4]   2> 3956365 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3956366 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3956367 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3956368 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3956367 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3956367 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3957202 T8042 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node4, state: recovering, checkLive: true, onlyIfLeader: true for: 1 seconds.
   [junit4]   2> 3957203 T8042 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node4&state=recovering&nodeName=127.0.0.1:51265_qb&action=PREPRECOVERY&checkLive=true&core=collection1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=1002 
   [junit4]   2> 3957207 T7990 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3958211 T7990 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2>  C1816_STATE=coll:collection1 core:collection1 props:{state=recovering, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3959206 T8085 C1816 P51265 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:51258/qb/collection1/ core=collection1 - recoveringAfterStartup=true
   [junit4]   2> 3959208 T8085 C1816 P51265 oasu.PeerSync.sync PeerSync: core=collection1 url=http://127.0.0.1:51265/qb START replicas=[http://127.0.0.1:51258/qb/collection1/] nUpdates=100
   [junit4]   2> 3959214 T8085 C1816 P51265 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 3959215 T8085 C1816 P51265 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=collection1
   [junit4]   2> 3959215 T8085 C1816 P51265 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=collection1
   [junit4]   2> 3959215 T7990 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3959215 T8085 C1816 P51265 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=collection1
   [junit4]   2> 3959217 T8085 C1816 P51265 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 3959217 T8085 C1816 P51265 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:51258/qb/collection1/. core=collection1
   [junit4]   2> 3959218 T8085 C1816 P51265 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> ASYNC  NEW_CORE C1817 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1817_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3959233 T8039 C1817 P51258 oasc.SolrCore.execute [collection1] webapp=/qb path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=3 
   [junit4]   2> 3959270 T8040 C1817 P51258 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3959284 T8040 C1817 P51258 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_1,generation=1}
   [junit4]   2> 3959286 T8040 C1817 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3959286 T8040 C1817 P51258 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 3959287 T8040 C1817 P51258 oass.SolrIndexSearcher.<init> Opening Searcher@3aee92b1 realtime
   [junit4]   2> 3959287 T8040 C1817 P51258 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3959288 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 18
   [junit4]   2> 3959292 T8085 C1816 P51265 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 3959293 T8085 C1816 P51265 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 3959297 T8043 C1817 P51258 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 3959298 T8043 C1817 P51258 oasc.SolrCore.execute [collection1] webapp=/qb path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=2 
   [junit4]   2> 3959299 T8085 C1816 P51265 oasc.RecoveryStrategy.replay No replay needed. core=collection1
   [junit4]   2> 3959299 T8085 C1816 P51265 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=collection1
   [junit4]   2> 3959300 T8085 C1816 P51265 oasc.ZkController.publish publishing core=collection1 state=active
   [junit4]   2> 3959300 T8085 C1816 P51265 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3959303 T8085 C1816 P51265 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=collection1
   [junit4]   2> 3959391 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3959393 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node4",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51265_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51265/qb"}
   [junit4]   2> 3959402 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3959403 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3959403 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3959403 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3959404 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3959404 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3960220 T7990 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3960222 T7990 oasc.AbstractDistribZkTestBase.waitForRecoveriesToFinish Recoveries finished - collection: collection1
   [junit4]   2> ASYNC  NEW_CORE C1818 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1818_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3960259 T8007 C1818 P51251 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3960265 T8007 C1818 P51251 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@65e17bbd lockFactory=org.apache.lucene.store.NativeFSLockFactory@7c5b2df5),segFN=segments_1,generation=1}
   [junit4]   2> 3960267 T8007 C1818 P51251 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3960269 T8007 C1818 P51251 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 3960269 T8007 C1818 P51251 oass.SolrIndexSearcher.<init> Opening Searcher@3a3126a main
   [junit4]   2> 3960270 T8007 C1818 P51251 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3960271 T8016 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@3a3126a main{StandardDirectoryReader(segments_1:1:nrt)}
   [junit4]   2> 3960278 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&version=2&softCommit=false} {commit=} 0 20
   [junit4]   2> 3960295 T8039 C1817 P51258 oasu.SolrCmdDistributor.distribCommit Distrib commit to:[StdNode: http://127.0.0.1:51255/qb/collection1/, StdNode: http://127.0.0.1:51261/qb/collection1/, StdNode: http://127.0.0.1:51258/qb/collection1/, StdNode: http://127.0.0.1:51265/qb/collection1/] params:commit_end_point=true&commit=true&softCommit=false&waitSearcher=true&expungeDeletes=false
   [junit4]   2> ASYNC  NEW_CORE C1819 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1819_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3960330 T8023 C1819 P51255 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3960331 T8023 C1819 P51255 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 3960332 T8023 C1819 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@62fef0f1 main
   [junit4]   2> 3960332 T8023 C1819 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3960334 T8040 C1817 P51258 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3960334 T8040 C1817 P51258 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 3960334 T8040 C1817 P51258 oass.SolrIndexSearcher.<init> Opening Searcher@7b12facf main
   [junit4]   2> 3960335 T8040 C1817 P51258 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3960337 T8034 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@62fef0f1 main{StandardDirectoryReader(segments_1:1:nrt)}
   [junit4]   2> 3960338 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 8
   [junit4]   2> ASYNC  NEW_CORE C1820 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1820_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3960338 T8053 C1820 P51261 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3960340 T8050 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@7b12facf main{StandardDirectoryReader(segments_1:1:nrt)}
   [junit4]   2> 3960340 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 7
   [junit4]   2> 3960343 T8053 C1820 P51261 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@4cf6277a lockFactory=org.apache.lucene.store.NativeFSLockFactory@518a7bdf),segFN=segments_1,generation=1}
   [junit4]   2> 3960346 T8053 C1820 P51261 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3960346 T8053 C1820 P51261 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 3960347 T8053 C1820 P51261 oass.SolrIndexSearcher.<init> Opening Searcher@342cd0e7 main
   [junit4]   2> 3960348 T8053 C1820 P51261 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2>  C1816_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3960349 T8073 C1816 P51265 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3960350 T8066 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@342cd0e7 main{StandardDirectoryReader(segments_1:1:nrt)}
   [junit4]   2> 3960350 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 12
   [junit4]   2> 3960353 T8073 C1816 P51265 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191fa213 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1b45ecd0),segFN=segments_1,generation=1}
   [junit4]   2> 3960355 T8073 C1816 P51265 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3960355 T8073 C1816 P51265 oasu.DirectUpdateHandler2.commit No uncommitted changes. Skipping IW.commit.
   [junit4]   2> 3960356 T8073 C1816 P51265 oass.SolrIndexSearcher.<init> Opening Searcher@62a7c2ac main
   [junit4]   2> 3960357 T8073 C1816 P51265 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3960358 T8084 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@62a7c2ac main{StandardDirectoryReader(segments_1:1:nrt)}
   [junit4]   2> 3960359 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 10
   [junit4]   2> 3960360 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&version=2&softCommit=false} {commit=} 0 65
   [junit4]   2> 3960362 T7990 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3960367 T8024 C1819 P51255 oasc.SolrCore.execute [collection1] webapp=/qb path=/select params={tests=checkShardConsistency&q=*:*&distrib=false&wt=javabin&version=2&rows=0} hits=0 status=0 QTime=1 
   [junit4]   2> 3960370 T8055 C1820 P51261 oasc.SolrCore.execute [collection1] webapp=/qb path=/select params={tests=checkShardConsistency&q=*:*&distrib=false&wt=javabin&version=2&rows=0} hits=0 status=0 QTime=0 
   [junit4]   2> 3960375 T8041 C1817 P51258 oasc.SolrCore.execute [collection1] webapp=/qb path=/select params={tests=checkShardConsistency&q=*:*&distrib=false&wt=javabin&version=2&rows=0} hits=0 status=0 QTime=1 
   [junit4]   2> 3960378 T8074 C1816 P51265 oasc.SolrCore.execute [collection1] webapp=/qb path=/select params={tests=checkShardConsistency&q=*:*&distrib=false&wt=javabin&version=2&rows=0} hits=0 status=0 QTime=0 
   [junit4]   2> 3962386 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {deleteByQuery=*:* (-1440564851064700928)} 0 3
   [junit4]   2> 3962414 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=FROMLEADER&_version_=-1440564851073089536&update.from=http://127.0.0.1:51258/qb/collection1/&wt=javabin&version=2} {deleteByQuery=*:* (-1440564851073089536)} 0 8
   [junit4]   2> 3962422 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=FROMLEADER&_version_=-1440564851081478144&update.from=http://127.0.0.1:51255/qb/collection1/&wt=javabin&version=2} {deleteByQuery=*:* (-1440564851081478144)} 0 4
   [junit4]   2> 3962423 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {deleteByQuery=*:* (-1440564851081478144)} 0 25
   [junit4]   2> 3962425 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {deleteByQuery=*:* (-1440564851073089536)} 0 34
   [junit4]   2> 3962431 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[a!0 (1440564851113984000)]} 0 2
   [junit4]   2> 3962452 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[a!0 (1440564851121324032)]} 0 8
   [junit4]   2> 3962454 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[a!0 (1440564851121324032)]} 0 18
   [junit4]   2> 3962484 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[b!1 (1440564851169558528)]} 0 1
   [junit4]   2> 3962517 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[b!1 (1440564851188432896)]} 0 13
   [junit4]   2> 3962519 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[b!1 (1440564851188432896)]} 0 19
   [junit4]   2> 3962546 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[c!2 (1440564851234570240)]} 0 1
   [junit4]   2> 3962563 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[c!2 (1440564851248201728)]} 0 1
   [junit4]   2> 3962566 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[c!2 (1440564851248201728)]} 0 8
   [junit4]   2> 3962579 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[d!3 (1440564851269173248)]} 0 1
   [junit4]   2> 3962594 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[d!3 (1440564851282804736)]} 0 1
   [junit4]   2> 3962596 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[d!3 (1440564851282804736)]} 0 6
   [junit4]   2> 3962597 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[d!3]} 0 10
   [junit4]   2> 3962603 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[e!4 (1440564851294339072)]} 0 1
   [junit4]   2> 3962639 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[e!4 (1440564851320553472)]} 0 1
   [junit4]   2> 3962647 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[e!4 (1440564851320553472)]} 0 20
   [junit4]   2> 3962651 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[e!4]} 0 35
   [junit4]   2> 3962663 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[f!5 (1440564851357253632)]} 0 1
   [junit4]   2> 3962691 T8072 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[f!5 (1440564851379273728)]} 0 1
   [junit4]   2> 3962693 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[f!5 (1440564851379273728)]} 0 10
   [junit4]   2> 3962694 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[f!5]} 0 25
   [junit4]   2> 3962698 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[g!6 (1440564851395002368)]} 0 0
   [junit4]   2> 3962706 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[g!6 (1440564851399196672)]} 0 1
   [junit4]   2> 3962711 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[g!6 (1440564851399196672)]} 0 9
   [junit4]   2> 3962719 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[h!7 (1440564851415973888)]} 0 1
   [junit4]   2> 3962746 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[h!7 (1440564851429605376)]} 0 4
   [junit4]   2> 3962753 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[h!7 (1440564851429605376)]} 0 23
   [junit4]   2> 3962762 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[i!8 (1440564851461062656)]} 0 1
   [junit4]   2> 3962790 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[i!8 (1440564851472596992)]} 0 0
   [junit4]   2> 3962792 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[i!8 (1440564851472596992)]} 0 21
   [junit4]   2> 3962797 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[j!9 (1440564851497762816)]} 0 1
   [junit4]   2> 3962829 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[j!9 (1440564851516637184)]} 0 5
   [junit4]   2> 3962832 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[j!9 (1440564851516637184)]} 0 18
   [junit4]   2> 3962849 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[k!10 (1440564851552288768)]} 0 1
   [junit4]   2> 3962875 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[k!10 (1440564851576406016)]} 0 1
   [junit4]   2> 3962876 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[k!10 (1440564851576406016)]} 0 5
   [junit4]   2> 3962881 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[l!11 (1440564851586891776)]} 0 1
   [junit4]   2> 3962902 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[l!11 (1440564851601571840)]} 0 2
   [junit4]   2> 3962903 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[l!11 (1440564851601571840)]} 0 8
   [junit4]   2> 3962914 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[m!12 (1440564851621494784)]} 0 1
   [junit4]   2> 3962929 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[m!12 (1440564851627786240)]} 0 5
   [junit4]   2> 3962930 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[m!12 (1440564851627786240)]} 0 10
   [junit4]   2> 3962935 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[n!13 (1440564851642466304)]} 0 1
   [junit4]   2> 3962956 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[n!13 (1440564851659243520)]} 0 2
   [junit4]   2> 3962960 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[n!13 (1440564851659243520)]} 0 11
   [junit4]   2> 3962964 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[o!14 (1440564851673923584)]} 0 0
   [junit4]   2> 3962987 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[o!14 (1440564851692797952)]} 0 1
   [junit4]   2> 3962988 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[o!14 (1440564851692797952)]} 0 6
   [junit4]   2> 3962995 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[p!15 (1440564851705380864)]} 0 1
   [junit4]   2> 3963011 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[p!15 (1440564851719012352)]} 0 1
   [junit4]   2> 3963013 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[p!15 (1440564851719012352)]} 0 6
   [junit4]   2> 3963014 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[p!15]} 0 11
   [junit4]   2> 3963018 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[q!16 (1440564851729498112)]} 0 1
   [junit4]   2> 3963030 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[q!16 (1440564851738935296)]} 0 0
   [junit4]   2> 3963032 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[q!16 (1440564851738935296)]} 0 6
   [junit4]   2> 3963033 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[q!16]} 0 11
   [junit4]   2> 3963037 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[r!17 (1440564851750469632)]} 0 0
   [junit4]   2> 3963046 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[r!17 (1440564851754663936)]} 0 2
   [junit4]   2> 3963048 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[r!17 (1440564851754663936)]} 0 7
   [junit4]   2> 3963052 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[s!18 (1440564851765149696)]} 0 1
   [junit4]   2> 3963062 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[s!18 (1440564851770392576)]} 0 4
   [junit4]   2> 3963068 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[s!18 (1440564851770392576)]} 0 13
   [junit4]   2> 3963074 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[t!19 (1440564851788218368)]} 0 1
   [junit4]   2> 3963086 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[t!19 (1440564851796606976)]} 0 1
   [junit4]   2> 3963088 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[t!19 (1440564851796606976)]} 0 7
   [junit4]   2> 3963092 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[t!19]} 0 14
   [junit4]   2> 3963097 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[u!20 (1440564851813384192)]} 0 1
   [junit4]   2> 3963108 T8072 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[u!20 (1440564851820724224)]} 0 0
   [junit4]   2> 3963110 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[u!20 (1440564851820724224)]} 0 6
   [junit4]   2> 3963112 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[u!20]} 0 11
   [junit4]   2> 3963116 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[v!21 (1440564851832258560)]} 0 1
   [junit4]   2> 3963127 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[v!21 (1440564851840647168)]} 0 1
   [junit4]   2> 3963128 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[v!21 (1440564851840647168)]} 0 5
   [junit4]   2> 3963130 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[v!21]} 0 10
   [junit4]   2> 3963135 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[w!22 (1440564851853230080)]} 0 1
   [junit4]   2> 3963158 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[w!22 (1440564851870007296)]} 0 1
   [junit4]   2> 3963159 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[w!22 (1440564851870007296)]} 0 9
   [junit4]   2> 3963160 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[w!22]} 0 14
   [junit4]   2> 3963168 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[x!23 (1440564851886784512)]} 0 1
   [junit4]   2> 3963182 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[x!23 (1440564851898318848)]} 0 0
   [junit4]   2> 3963183 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[x!23 (1440564851898318848)]} 0 5
   [junit4]   2> 3963184 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[x!23]} 0 11
   [junit4]   2> 3963188 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[y!24 (1440564851907756032)]} 0 1
   [junit4]   2> 3963200 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[y!24 (1440564851917193216)]} 0 1
   [junit4]   2> 3963201 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[y!24 (1440564851917193216)]} 0 5
   [junit4]   2> 3963202 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[y!24]} 0 9
   [junit4]   2> 3963207 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[z!25 (1440564851927678976)]} 0 1
   [junit4]   2> 3963214 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[z!25 (1440564851932921856)]} 0 0
   [junit4]   2> 3963217 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[z!25 (1440564851932921856)]} 0 7
   [junit4]   2> 3963226 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[a!26 (1440564851947601920)]} 0 2
   [junit4]   2> 3963237 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[a!26 (1440564851955990528)]} 0 1
   [junit4]   2> 3963239 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[a!26 (1440564851955990528)]} 0 6
   [junit4]   2> 3963243 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[b!27 (1440564851965427712)]} 0 1
   [junit4]   2> 3963252 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[b!27 (1440564851972767744)]} 0 0
   [junit4]   2> 3963253 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[b!27 (1440564851972767744)]} 0 4
   [junit4]   2> 3963254 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[b!27]} 0 7
   [junit4]   2> 3963258 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[c!28 (1440564851981156352)]} 0 1
   [junit4]   2> 3963268 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[c!28 (1440564851987447808)]} 0 0
   [junit4]   2> 3963269 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[c!28 (1440564851987447808)]} 0 6
   [junit4]   2> 3963270 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[c!28]} 0 9
   [junit4]   2> 3963274 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[d!29 (1440564851998982144)]} 0 1
   [junit4]   2> 3963285 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[d!29 (1440564852002127872)]} 0 3
   [junit4]   2> 3963286 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[d!29 (1440564852002127872)]} 0 9
   [junit4]   2> 3963291 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[e!30 (1440564852016807936)]} 0 0
   [junit4]   2> 3963304 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[e!30 (1440564852024147968)]} 0 1
   [junit4]   2> 3963306 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[e!30 (1440564852024147968)]} 0 8
   [junit4]   2> 3963311 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[f!31 (1440564852036730880)]} 0 1
   [junit4]   2> 3963319 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[f!31 (1440564852041973760)]} 0 1
   [junit4]   2> 3963322 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[f!31 (1440564852041973760)]} 0 8
   [junit4]   2> 3963325 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[g!32 (1440564852052459520)]} 0 1
   [junit4]   2> 3963332 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[g!32 (1440564852056653824)]} 0 1
   [junit4]   2> 3963334 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[g!32 (1440564852056653824)]} 0 6
   [junit4]   2> 3963338 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[h!33 (1440564852066091008)]} 0 1
   [junit4]   2> 3963345 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[h!33 (1440564852069236736)]} 0 1
   [junit4]   2> 3963347 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[h!33 (1440564852069236736)]} 0 6
   [junit4]   2> 3963350 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[i!34 (1440564852078673920)]} 0 0
   [junit4]   2> 3963363 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[i!34 (1440564852084965376)]} 0 1
   [junit4]   2> 3963365 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[i!34 (1440564852084965376)]} 0 10
   [junit4]   2> 3963370 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[j!35 (1440564852099645440)]} 0 0
   [junit4]   2> 3963379 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[j!35 (1440564852104888320)]} 0 1
   [junit4]   2> 3963380 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[j!35 (1440564852104888320)]} 0 5
   [junit4]   2> 3963385 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[k!36 (1440564852114325504)]} 0 1
   [junit4]   2> 3963395 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[k!36 (1440564852121665536)]} 0 1
   [junit4]   2> 3963396 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[k!36 (1440564852121665536)]} 0 5
   [junit4]   2> 3963397 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[k!36]} 0 9
   [junit4]   2> 3963402 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[l!37 (1440564852132151296)]} 0 1
   [junit4]   2> 3963409 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[l!37 (1440564852136345600)]} 0 1
   [junit4]   2> 3963410 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[l!37 (1440564852136345600)]} 0 5
   [junit4]   2> 3963415 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[m!38 (1440564852144734208)]} 0 2
   [junit4]   2> 3963428 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[m!38 (1440564852155219968)]} 0 1
   [junit4]   2> 3963429 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[m!38 (1440564852155219968)]} 0 6
   [junit4]   2> 3963433 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[n!39 (1440564852164657152)]} 0 1
   [junit4]   2> 3963442 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[n!39 (1440564852169900032)]} 0 2
   [junit4]   2> 3963445 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[n!39 (1440564852169900032)]} 0 8
   [junit4]   2> 3963451 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[o!40 (1440564852183531520)]} 0 2
   [junit4]   2> 3963463 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[o!40 (1440564852194017280)]} 0 1
   [junit4]   2> 3963464 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[o!40 (1440564852194017280)]} 0 4
   [junit4]   2> 3963466 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[o!40]} 0 9
   [junit4]   2> 3963470 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[p!41 (1440564852203454464)]} 0 1
   [junit4]   2> 3963479 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[p!41 (1440564852210794496)]} 0 1
   [junit4]   2> 3963480 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[p!41 (1440564852210794496)]} 0 4
   [junit4]   2> 3963482 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[p!41]} 0 9
   [junit4]   2> 3963489 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[q!42 (1440564852223377408)]} 0 2
   [junit4]   2> 3963509 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[q!42 (1440564852241203200)]} 0 1
   [junit4]   2> 3963511 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[q!42 (1440564852241203200)]} 0 7
   [junit4]   2> 3963516 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[r!43 (1440564852251688960)]} 0 1
   [junit4]   2> 3963526 T8072 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[r!43 (1440564852259028992)]} 0 1
   [junit4]   2> 3963528 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[r!43 (1440564852259028992)]} 0 7
   [junit4]   2> 3963529 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[r!43]} 0 10
   [junit4]   2> 3963534 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[s!44 (1440564852270563328)]} 0 1
   [junit4]   2> 3963541 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[s!44 (1440564852275806208)]} 0 1
   [junit4]   2> 3963547 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[s!44 (1440564852275806208)]} 0 10
   [junit4]   2> 3963554 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[t!45 (1440564852292583424)]} 0 1
   [junit4]   2> 3963568 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[t!45 (1440564852303069184)]} 0 1
   [junit4]   2> 3963569 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[t!45 (1440564852303069184)]} 0 6
   [junit4]   2> 3963574 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[u!46 (1440564852313554944)]} 0 1
   [junit4]   2> 3963582 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[u!46 (1440564852317749248)]} 0 1
   [junit4]   2> 3963583 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[u!46 (1440564852317749248)]} 0 5
   [junit4]   2> 3963587 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[v!47 (1440564852326137856)]} 0 1
   [junit4]   2> 3963596 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[v!47 (1440564852333477888)]} 0 1
   [junit4]   2> 3963598 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[v!47 (1440564852333477888)]} 0 5
   [junit4]   2> 3963599 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[v!47]} 0 9
   [junit4]   2> 3963604 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[w!48 (1440564852343963648)]} 0 1
   [junit4]   2> 3963621 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[w!48 (1440564852354449408)]} 0 1
   [junit4]   2> 3963623 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[w!48 (1440564852354449408)]} 0 11
   [junit4]   2> 3963624 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[w!48]} 0 16
   [junit4]   2> 3963630 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[x!49 (1440564852371226624)]} 0 2
   [junit4]   2> 3963640 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[x!49 (1440564852377518080)]} 0 1
   [junit4]   2> 3963642 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[x!49 (1440564852377518080)]} 0 7
   [junit4]   2> 3963646 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[y!50 (1440564852389052416)]} 0 1
   [junit4]   2> 3963657 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[y!50 (1440564852396392448)]} 0 1
   [junit4]   2> 3963659 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[y!50 (1440564852396392448)]} 0 6
   [junit4]   2> 3963660 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[y!50]} 0 10
   [junit4]   2> 3963664 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[z!51 (1440564852406878208)]} 0 1
   [junit4]   2> 3963673 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[z!51 (1440564852414218240)]} 0 0
   [junit4]   2> 3963675 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[z!51 (1440564852414218240)]} 0 5
   [junit4]   2> 3963676 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[z!51]} 0 9
   [junit4]   2> 3963680 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[a!52 (1440564852424704000)]} 0 1
   [junit4]   2> 3963689 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[a!52 (1440564852430995456)]} 0 0
   [junit4]   2> 3963691 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[a!52 (1440564852430995456)]} 0 5
   [junit4]   2> 3963692 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[a!52]} 0 8
   [junit4]   2> 3963701 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[b!53 (1440564852446724096)]} 0 1
   [junit4]   2> 3963714 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[b!53 (1440564852456161280)]} 0 1
   [junit4]   2> 3963715 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[b!53 (1440564852456161280)]} 0 6
   [junit4]   2> 3963717 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[b!53]} 0 12
   [junit4]   2> 3963721 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[c!54 (1440564852467695616)]} 0 0
   [junit4]   2> 3963728 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[c!54 (1440564852471889920)]} 0 0
   [junit4]   2> 3963730 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[c!54 (1440564852471889920)]} 0 5
   [junit4]   2> 3963734 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[d!55 (1440564852480278528)]} 0 1
   [junit4]   2> 3963741 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[d!55 (1440564852485521408)]} 0 1
   [junit4]   2> 3963744 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[d!55 (1440564852485521408)]} 0 7
   [junit4]   2> 3963752 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[e!56 (1440564852500201472)]} 0 1
   [junit4]   2> 3963763 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[e!56 (1440564852506492928)]} 0 1
   [junit4]   2> 3963764 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[e!56 (1440564852506492928)]} 0 6
   [junit4]   2> 3963768 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[f!57 (1440564852515930112)]} 0 1
   [junit4]   2> 3963776 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[f!57 (1440564852521172992)]} 0 1
   [junit4]   2> 3963778 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[f!57 (1440564852521172992)]} 0 6
   [junit4]   2> 3963782 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[g!58 (1440564852531658752)]} 0 1
   [junit4]   2> 3963789 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[g!58 (1440564852535853056)]} 0 0
   [junit4]   2> 3963791 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[g!58 (1440564852535853056)]} 0 5
   [junit4]   2> 3963795 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[h!59 (1440564852544241664)]} 0 1
   [junit4]   2> 3963803 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[h!59 (1440564852549484544)]} 0 1
   [junit4]   2> 3963804 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[h!59 (1440564852549484544)]} 0 5
   [junit4]   2> 3963808 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[i!60 (1440564852558921728)]} 0 1
   [junit4]   2> 3963822 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[i!60 (1440564852565213184)]} 0 1
   [junit4]   2> 3963824 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[i!60 (1440564852565213184)]} 0 12
   [junit4]   2> 3963835 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[j!61 (1440564852586184704)]} 0 1
   [junit4]   2> 3963844 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[j!61 (1440564852593524736)]} 0 0
   [junit4]   2> 3963846 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[j!61 (1440564852593524736)]} 0 5
   [junit4]   2> 3963847 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[j!61]} 0 9
   [junit4]   2> 3963852 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[k!62 (1440564852604010496)]} 0 1
   [junit4]   2> 3963859 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[k!62 (1440564852608204800)]} 0 1
   [junit4]   2> 3963860 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[k!62 (1440564852608204800)]} 0 5
   [junit4]   2> 3963864 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[l!63 (1440564852617641984)]} 0 0
   [junit4]   2> 3963871 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[l!63 (1440564852621836288)]} 0 1
   [junit4]   2> 3963873 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[l!63 (1440564852621836288)]} 0 5
   [junit4]   2> 3963882 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[m!64 (1440564852632322048)]} 0 5
   [junit4]   2> 3963889 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[m!64 (1440564852640710656)]} 0 1
   [junit4]   2> 3963891 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[m!64 (1440564852640710656)]} 0 5
   [junit4]   2> 3963894 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[n!65 (1440564852649099264)]} 0 0
   [junit4]   2> 3963903 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[n!65 (1440564852654342144)]} 0 1
   [junit4]   2> 3963904 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[n!65 (1440564852654342144)]} 0 5
   [junit4]   2> 3963910 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[o!66 (1440564852664827904)]} 0 2
   [junit4]   2> 3963920 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[o!66 (1440564852671119360)]} 0 1
   [junit4]   2> 3963922 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[o!66 (1440564852671119360)]} 0 7
   [junit4]   2> 3963927 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[p!67 (1440564852682653696)]} 0 1
   [junit4]   2> 3963935 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[p!67 (1440564852687896576)]} 0 1
   [junit4]   2> 3963936 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[p!67 (1440564852687896576)]} 0 6
   [junit4]   2> 3963940 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[q!68 (1440564852696285184)]} 0 1
   [junit4]   2> 3963952 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[q!68 (1440564852701528064)]} 0 3
   [junit4]   2> 3963954 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[q!68 (1440564852701528064)]} 0 10
   [junit4]   2> 3963960 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[r!69 (1440564852717256704)]} 0 2
   [junit4]   2> 3963969 T8072 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[r!69 (1440564852722499584)]} 0 1
   [junit4]   2> 3963970 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[r!69 (1440564852722499584)]} 0 6
   [junit4]   2> 3963975 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[s!70 (1440564852732985344)]} 0 1
   [junit4]   2> 3963987 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[s!70 (1440564852742422528)]} 0 1
   [junit4]   2> 3963989 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[s!70 (1440564852742422528)]} 0 7
   [junit4]   2> 3963990 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[s!70]} 0 11
   [junit4]   2> 3963996 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[t!71 (1440564852753956864)]} 0 2
   [junit4]   2> 3964002 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[t!71 (1440564852759199744)]} 0 0
   [junit4]   2> 3964004 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[t!71 (1440564852759199744)]} 0 5
   [junit4]   2> 3964009 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[u!72 (1440564852769685504)]} 0 1
   [junit4]   2> 3964024 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[u!72 (1440564852778074112)]} 0 1
   [junit4]   2> 3964025 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[u!72 (1440564852778074112)]} 0 8
   [junit4]   2> 3964030 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[v!73 (1440564852790657024)]} 0 1
   [junit4]   2> 3964037 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[v!73 (1440564852795899904)]} 0 1
   [junit4]   2> 3964039 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[v!73 (1440564852795899904)]} 0 6
   [junit4]   2> 3964042 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[w!74 (1440564852804288512)]} 0 0
   [junit4]   2> 3964049 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[w!74 (1440564852807434240)]} 0 1
   [junit4]   2> 3964051 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[w!74 (1440564852807434240)]} 0 6
   [junit4]   2> 3964055 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[x!75 (1440564852817920000)]} 0 0
   [junit4]   2> 3964064 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[x!75 (1440564852824211456)]} 0 1
   [junit4]   2> 3964066 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[x!75 (1440564852824211456)]} 0 5
   [junit4]   2> 3964067 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[x!75]} 0 9
   [junit4]   2> 3964070 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[y!76 (1440564852833648640)]} 0 0
   [junit4]   2> 3964076 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[y!76 (1440564852836794368)]} 0 0
   [junit4]   2> 3964077 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[y!76 (1440564852836794368)]} 0 4
   [junit4]   2> 3964081 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[z!77 (1440564852845182976)]} 0 0
   [junit4]   2> 3964088 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[z!77 (1440564852849377280)]} 0 1
   [junit4]   2> 3964090 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[z!77 (1440564852849377280)]} 0 5
   [junit4]   2> 3964097 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[a!78 (1440564852860911616)]} 0 2
   [junit4]   2> 3964108 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[a!78 (1440564852869300224)]} 0 1
   [junit4]   2> 3964109 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[a!78 (1440564852869300224)]} 0 6
   [junit4]   2> 3964114 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[b!79 (1440564852878737408)]} 0 1
   [junit4]   2> 3964121 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[b!79 (1440564852882931712)]} 0 1
   [junit4]   2> 3964123 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[b!79 (1440564852882931712)]} 0 6
   [junit4]   2> 3964127 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[c!80 (1440564852892368896)]} 0 1
   [junit4]   2> 3964136 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[c!80 (1440564852899708928)]} 0 1
   [junit4]   2> 3964138 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[c!80 (1440564852899708928)]} 0 6
   [junit4]   2> 3964139 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[c!80]} 0 9
   [junit4]   2> 3964143 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[d!81 (1440564852909146112)]} 0 1
   [junit4]   2> 3964170 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[d!81 (1440564852921729024)]} 0 7
   [junit4]   2> 3964173 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[d!81 (1440564852921729024)]} 0 19
   [junit4]   2> 3964177 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[e!82 (1440564852944797696)]} 0 1
   [junit4]   2> 3964190 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[e!82 (1440564852950040576)]} 0 1
   [junit4]   2> 3964191 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[e!82 (1440564852950040576)]} 0 10
   [junit4]   2> 3964195 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[f!83 (1440564852964720640)]} 0 1
   [junit4]   2> 3964205 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[f!83 (1440564852971012096)]} 0 0
   [junit4]   2> 3964207 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[f!83 (1440564852971012096)]} 0 6
   [junit4]   2> 3964209 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[f!83]} 0 11
   [junit4]   2> 3964213 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[g!84 (1440564852982546432)]} 0 1
   [junit4]   2> 3964223 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[g!84 (1440564852989886464)]} 0 1
   [junit4]   2> 3964224 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[g!84 (1440564852989886464)]} 0 5
   [junit4]   2> 3964226 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[g!84]} 0 10
   [junit4]   2> 3964230 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[h!85 (1440564853000372224)]} 0 1
   [junit4]   2> 3964239 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[h!85 (1440564853006663680)]} 0 1
   [junit4]   2> 3964240 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[h!85 (1440564853006663680)]} 0 5
   [junit4]   2> 3964242 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[h!85]} 0 9
   [junit4]   2> 3964246 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[i!86 (1440564853018198016)]} 0 1
   [junit4]   2> 3964260 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[i!86 (1440564853024489472)]} 0 1
   [junit4]   2> 3964261 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[i!86 (1440564853024489472)]} 0 9
   [junit4]   2> 3964262 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[i!86]} 0 13
   [junit4]   2> 3964266 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[j!87 (1440564853038120960)]} 0 1
   [junit4]   2> 3964277 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[j!87 (1440564853045460992)]} 0 1
   [junit4]   2> 3964279 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[j!87 (1440564853045460992)]} 0 8
   [junit4]   2> 3964281 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[j!87]} 0 13
   [junit4]   2> 3964285 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[k!88 (1440564853059092480)]} 0 1
   [junit4]   2> 3964295 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[k!88 (1440564853065383936)]} 0 1
   [junit4]   2> 3964296 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[k!88 (1440564853065383936)]} 0 5
   [junit4]   2> 3964297 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[k!88]} 0 9
   [junit4]   2> 3964302 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[l!89 (1440564853075869696)]} 0 1
   [junit4]   2> 3964312 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[l!89 (1440564853083209728)]} 0 0
   [junit4]   2> 3964318 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[l!89 (1440564853083209728)]} 0 11
   [junit4]   2> 3964320 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[l!89]} 0 15
   [junit4]   2> 3964328 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[m!90 (1440564853104181248)]} 0 1
   [junit4]   2> 3964339 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[m!90 (1440564853108375552)]} 0 1
   [junit4]   2> 3964340 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[m!90 (1440564853108375552)]} 0 8
   [junit4]   2> 3964345 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[n!91 (1440564853120958464)]} 0 1
   [junit4]   2> 3964355 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[n!91 (1440564853128298496)]} 0 1
   [junit4]   2> 3964357 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[n!91 (1440564853128298496)]} 0 6
   [junit4]   2> 3964358 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[n!91]} 0 9
   [junit4]   2> 3964362 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[o!92 (1440564853138784256)]} 0 1
   [junit4]   2> 3964369 T8072 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[o!92 (1440564853144027136)]} 0 0
   [junit4]   2> 3964371 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[o!92 (1440564853144027136)]} 0 5
   [junit4]   2> 3964375 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[p!93 (1440564853152415744)]} 0 1
   [junit4]   2> 3964387 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[p!93 (1440564853157658624)]} 0 3
   [junit4]   2> 3964389 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[p!93 (1440564853157658624)]} 0 11
   [junit4]   2> 3964393 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[q!94 (1440564853172338688)]} 0 1
   [junit4]   2> 3964406 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[q!94 (1440564853182824448)]} 0 0
   [junit4]   2> 3964410 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[q!94 (1440564853182824448)]} 0 8
   [junit4]   2> 3964412 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[q!94]} 0 14
   [junit4]   2> 3964417 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[r!95 (1440564853196455936)]} 0 1
   [junit4]   2> 3964428 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[r!95 (1440564853203795968)]} 0 1
   [junit4]   2> 3964430 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[r!95 (1440564853203795968)]} 0 7
   [junit4]   2> 3964433 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[s!96 (1440564853214281728)]} 0 0
   [junit4]   2> 3964441 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[s!96 (1440564853218476032)]} 0 1
   [junit4]   2> 3964443 T8027 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[s!96 (1440564853218476032)]} 0 6
   [junit4]   2> 3964447 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[t!97 (1440564853227913216)]} 0 1
   [junit4]   2> 3964462 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[t!97 (1440564853239447552)]} 0 0
   [junit4]   2> 3964464 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[t!97 (1440564853239447552)]} 0 7
   [junit4]   2> 3964465 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[t!97]} 0 14
   [junit4]   2> 3964469 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[u!98 (1440564853252030464)]} 0 1
   [junit4]   2> 3964480 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[u!98 (1440564853260419072)]} 0 0
   [junit4]   2> 3964482 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[u!98 (1440564853260419072)]} 0 5
   [junit4]   2> 3964484 T8028 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[u!98]} 0 10
   [junit4]   2> 3964489 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[v!99 (1440564853271953408)]} 0 1
   [junit4]   2> 3964499 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[v!99 (1440564853279293440)]} 0 1
   [junit4]   2> 3964500 T8023 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[v!99 (1440564853279293440)]} 0 5
   [junit4]   2> 3964502 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[v!99]} 0 10
   [junit4]   2> 3964506 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[w!100 (1440564853289779200)]} 0 1
   [junit4]   2> 3964522 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[w!100 (1440564853298167808)]} 0 1
   [junit4]   2> 3964523 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[w!100 (1440564853298167808)]} 0 11
   [junit4]   2> 3964524 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[w!100]} 0 15
   [junit4]   2> 3964528 T8007 C1818 P51251 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3964531 T8007 C1818 P51251 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@65e17bbd lockFactory=org.apache.lucene.store.NativeFSLockFactory@7c5b2df5),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@65e17bbd lockFactory=org.apache.lucene.store.NativeFSLockFactory@7c5b2df5),segFN=segments_2,generation=2}
   [junit4]   2> 3964532 T8007 C1818 P51251 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 2
   [junit4]   2> 3964532 T8007 C1818 P51251 oass.SolrIndexSearcher.<init> Opening Searcher@3eb66a1d main
   [junit4]   2> 3964533 T8007 C1818 P51251 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3964534 T8016 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@3eb66a1d main{StandardDirectoryReader(segments_2:3:nrt _0(5.0):C101)}
   [junit4]   2> 3964535 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&version=2&softCommit=false} {commit=} 0 7
   [junit4]   2> 3964539 T8042 C1817 P51258 oasu.SolrCmdDistributor.distribCommit Distrib commit to:[StdNode: http://127.0.0.1:51255/qb/collection1/, StdNode: http://127.0.0.1:51261/qb/collection1/, StdNode: http://127.0.0.1:51258/qb/collection1/, StdNode: http://127.0.0.1:51265/qb/collection1/] params:commit_end_point=true&commit=true&softCommit=false&waitSearcher=true&expungeDeletes=false
   [junit4]   2> 3964542 T8074 C1816 P51265 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3964542 T8025 C1819 P51255 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3964543 T8044 C1817 P51258 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3964544 T8055 C1820 P51261 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3964549 T8074 C1816 P51265 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191fa213 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1b45ecd0),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191fa213 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1b45ecd0),segFN=segments_2,generation=2}
   [junit4]   2> 3964550 T8074 C1816 P51265 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 2
   [junit4]   2> 3964549 T8044 C1817 P51258 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_2,generation=2}
   [junit4]   2> 3964552 T8044 C1817 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 2
   [junit4]   2> 3964551 T8074 C1816 P51265 oass.SolrIndexSearcher.<init> Opening Searcher@259dbda8 main
   [junit4]   2> 3964553 T8074 C1816 P51265 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3964555 T8084 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@259dbda8 main{StandardDirectoryReader(segments_2:3:nrt _0(5.0):C42)}
   [junit4]   2> 3964556 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 14
   [junit4]   2> 3964559 T8044 C1817 P51258 oass.SolrIndexSearcher.<init> Opening Searcher@3ee382fe main
   [junit4]   2> 3964561 T8044 C1817 P51258 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3964563 T8050 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@3ee382fe main{StandardDirectoryReader(segments_2:3:nrt _0(5.0):C42)}
   [junit4]   2> 3964565 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 22
   [junit4]   2> 3964565 T8025 C1819 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_2,generation=2}
   [junit4]   2> 3964566 T8025 C1819 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 2
   [junit4]   2> 3964568 T8025 C1819 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@52f3d08b main
   [junit4]   2> 3964569 T8025 C1819 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3964571 T8034 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@52f3d08b main{StandardDirectoryReader(segments_2:3:nrt _0(5.0):C59)}
   [junit4]   2> 3964572 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 30
   [junit4]   2> 3964575 T8055 C1820 P51261 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@4cf6277a lockFactory=org.apache.lucene.store.NativeFSLockFactory@518a7bdf),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@4cf6277a lockFactory=org.apache.lucene.store.NativeFSLockFactory@518a7bdf),segFN=segments_2,generation=2}
   [junit4]   2> 3964576 T8055 C1820 P51261 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 2
   [junit4]   2> 3964577 T8055 C1820 P51261 oass.SolrIndexSearcher.<init> Opening Searcher@6fac5104 main
   [junit4]   2> 3964578 T8055 C1820 P51261 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3964579 T8066 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@6fac5104 main{StandardDirectoryReader(segments_2:3:nrt _0(5.0):C59)}
   [junit4]   2> 3964580 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 36
   [junit4]   2> 3964582 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&version=2&softCommit=false} {commit=} 0 44
   [junit4]   2> 3964585 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3964597 T8093 oasc.ShardSplitTest$1.run SHARDSPLITTEST: Going to add 518 number of docs at 1 doc per 28ms
   [junit4]   2> 3964614 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[101 (1440564853400928256)]} 0 3
   [junit4]   2> 3964615 T8027 oasha.CollectionsHandler.handleSplitShardAction Splitting shard : shard=shard1&action=SPLITSHARD&collection=collection1&wt=javabin&version=2
   [junit4]   2> 3964621 T8012 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work state: SyncConnected type NodeChildrenChanged
   [junit4]   2> 3964628 T8014 oasc.OverseerCollectionProcessor.run Overseer Collection Processor: Get the message id:/overseer/collection-queue-work/qn-0000000000 message:{
   [junit4]   2> 	  "operation":"splitshard",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "collection":"collection1"}
   [junit4]   2> 3964628 T8014 oasc.OverseerCollectionProcessor.splitShard Split shard invoked
   [junit4]   2> 3964629 T8014 oasc.OverseerCollectionProcessor.splitShard Creating shard collection1_shard1_0_replica1 as part of slice shard1_0 of collection collection1 on 127.0.0.1:51255_qb
   [junit4]   2> 3964630 T8014 oasc.OverseerCollectionProcessor.splitShard Creating shard collection1_shard1_1_replica1 as part of slice shard1_1 of collection collection1 on 127.0.0.1:51255_qb
   [junit4]   2> 3964630 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[101 (1440564853412462592)]} 0 2
   [junit4]   2> 3964634 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[101 (1440564853412462592)]} 0 13
   [junit4]   2> 3964642 T8028 oasc.ZkController.publish publishing core=collection1_shard1_0_replica1 state=down
   [junit4]   2> 3964643 T8028 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3964643 T8023 oasc.ZkController.publish publishing core=collection1_shard1_1_replica1 state=down
   [junit4]   2> 3964644 T8023 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3964648 T8028 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 3964649 T8023 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 3964668 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[101 (-1440564853460697088)]} 0 1
   [junit4]   2> 3964676 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[101 (-1440564853465939968)]} 0 1
   [junit4]   2> 3964677 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[101 (-1440564853465939968)]} 0 5
   [junit4]   2> 3964683 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[102 (1440564853475377152)]} 0 1
   [junit4]   2> 3964697 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[102 (1440564853485862912)]} 0 1
   [junit4]   2> 3964699 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[102 (1440564853485862912)]} 0 7
   [junit4]   2> 3964701 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[102]} 0 13
   [junit4]   2> 3964734 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[103 (1440564853529903104)]} 0 1
   [junit4]   2> 3964743 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[103 (1440564853533048832)]} 0 1
   [junit4]   2> 3964745 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[103 (1440564853533048832)]} 0 8
   [junit4]   2> 3964778 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[102 (-1440564853576040448)]} 0 0
   [junit4]   2> 3964791 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[102 (-1440564853581283328)]} 0 4
   [junit4]   2> 3964796 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[102 (-1440564853581283328)]} 0 13
   [junit4]   2> 3964801 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[104 (1440564853599109120)]} 0 1
   [junit4]   2> 3964811 T8072 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[104 (1440564853606449152)]} 0 1
   [junit4]   2> 3964813 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[104 (1440564853606449152)]} 0 6
   [junit4]   2> 3964814 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[104]} 0 10
   [junit4]   2> 3964851 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[105 (1440564853652586496)]} 0 1
   [junit4]   2> 3964862 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[105 (1440564853659926528)]} 0 1
   [junit4]   2> 3964863 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[105 (1440564853659926528)]} 0 5
   [junit4]   2> 3964896 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[104 (-1440564853699772416)]} 0 0
   [junit4]   2> 3964906 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[104 (-1440564853707112448)]} 0 1
   [junit4]   2> 3964907 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[104 (-1440564853707112448)]} 0 5
   [junit4]   2> 3964914 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[104]} 0 14
   [junit4]   2> 3964925 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[106 (1440564853729132544)]} 0 1
   [junit4]   2> 3964933 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[106 (1440564853734375424)]} 0 1
   [junit4]   2> 3964934 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[106 (1440564853734375424)]} 0 5
   [junit4]   2> 3964968 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[107 (1440564853774221312)]} 0 1
   [junit4]   2> 3964978 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[107 (1440564853781561344)]} 0 1
   [junit4]   2> 3964979 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[107 (1440564853781561344)]} 0 5
   [junit4]   2> 3964986 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[107]} 0 14
   [junit4]   2> 3965022 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[106 (-1440564853830844416)]} 0 1
   [junit4]   2> 3965029 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[106 (-1440564853837135872)]} 0 0
   [junit4]   2> 3965031 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[106 (-1440564853837135872)]} 0 5
   [junit4]   2> 3965035 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[108 (1440564853845524480)]} 0 1
   [junit4]   2> 3965053 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[108 (1440564853849718784)]} 0 8
   [junit4]   2> 3965054 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[108 (1440564853849718784)]} 0 15
   [junit4]   2> 3965088 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[108 (-1440564853901099008)]} 0 1
   [junit4]   2> 3965097 T8053 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[108 (-1440564853907390464)]} 0 0
   [junit4]   2> 3965099 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[108 (-1440564853907390464)]} 0 5
   [junit4]   2> 3965100 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[108]} 0 9
   [junit4]   2> 3965104 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[109 (1440564853916827648)]} 0 1
   [junit4]   2> 3965118 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[109 (1440564853921021952)]} 0 6
   [junit4]   2> 3965120 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[109 (1440564853921021952)]} 0 13
   [junit4]   2> 3965153 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[110 (1440564853968207872)]} 0 1
   [junit4]   2> 3965162 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[110 (1440564853975547904)]} 0 0
   [junit4]   2> 3965164 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[110 (1440564853975547904)]} 0 5
   [junit4]   2> 3965165 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[110]} 0 8
   [junit4]   2> 3965200 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[103 (-1440564854018539520)]} 0 0
   [junit4]   2> 3965213 T8072 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[103 (-1440564854022733824)]} 0 0
   [junit4]   2> 3965215 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[103 (-1440564854022733824)]} 0 11
   [junit4]   2> 3965220 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[111 (1440564854039511040)]} 0 0
   [junit4]   2> 3965230 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[111 (1440564854045802496)]} 0 0
   [junit4]   2> 3965232 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[111 (1440564854045802496)]} 0 6
   [junit4]   2> 3965233 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[111]} 0 10
   [junit4]   2> 3965271 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[112 (1440564854092988416)]} 0 1
   [junit4]   2> 3965284 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[112 (1440564854101377024)]} 0 1
   [junit4]   2> 3965293 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[112 (1440564854101377024)]} 0 14
   [junit4]   2> 3965295 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[112]} 0 19
   [junit4]   2> 3965328 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[113 (1440564854152757248)]} 0 1
   [junit4]   2> 3965342 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[113 (1440564854163243008)]} 0 1
   [junit4]   2> 3965343 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[113 (1440564854163243008)]} 0 5
   [junit4]   2> 3965345 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[113]} 0 10
   [junit4]   2> 3965384 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[114 (1440564854210428928)]} 0 1
   [junit4]   2> 3965395 T8058 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[114 (1440564854217768960)]} 0 0
   [junit4]   2> 3965397 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[114 (1440564854217768960)]} 0 7
   [junit4]   2> 3965398 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[114]} 0 11
   [junit4]   2> 3965430 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[111 (-1440564854259712000)]} 0 0
   [junit4]   2> 3965437 T8056 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[111 (-1440564854263906304)]} 0 0
   [junit4]   2> 3965438 T8025 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[111 (-1440564854263906304)]} 0 4
   [junit4]   2> 3965447 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[115 (1440564854274392064)]} 0 4
   [junit4]   2> 3965458 T8077 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[115 (1440564854281732096)]} 0 1
   [junit4]   2> 3965459 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[115 (1440564854281732096)]} 0 8
   [junit4]   2> 3965466 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3965469 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_0",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"construction",
   [junit4]   2> 	  "core":"collection1_shard1_0_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":"80000000-bfffffff",
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3965480 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"construction",
   [junit4]   2> 	  "core":"collection1_shard1_1_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":"c0000000-ffffffff",
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3965493 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3965494 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3965494 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3965495 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3965495 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3965495 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3965497 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[110 (-1440564854329966592)]} 0 0
   [junit4]   2> 3965509 T8057 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[110 (-1440564854337306624)]} 0 0
   [junit4]   2> 3965511 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[110 (-1440564854337306624)]} 0 7
   [junit4]   2> 3965515 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[116 (1440564854347792384)]} 0 1
   [junit4]   2> 3965530 T8073 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[116 (1440564854359326720)]} 0 1
   [junit4]   2> 3965531 T8042 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[116 (1440564854359326720)]} 0 7
   [junit4]   2> 3965565 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[117 (1440564854400221184)]} 0 1
   [junit4]   2> 3965573 T8075 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[117 (1440564854405464064)]} 0 1
   [junit4]   2> 3965575 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[117 (1440564854405464064)]} 0 6
   [junit4]   2> 3965616 T8005 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[114 (-1440564854454747136)]} 0 0
   [junit4]   2> 3965625 T8059 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[114 (-1440564854461038592)]} 0 0
   [junit4]   2> 3965627 T8026 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[114 (-1440564854461038592)]} 0 5
   [junit4]   2> 3965628 T8043 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[114]} 0 8
   [junit4]   2> 3965632 T8007 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[118 (1440564854470475776)]} 0 1
   [junit4]   2> 3965642 T8055 C1820 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[118 (1440564854477815808)]} 0 1
   [junit4]   2> 3965650 T8024 C1819 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[118 (1440564854477815808)]} 0 12
   [junit4]   2> 3965651 T8023 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1_shard1_1_replica1
   [junit4]   2> 3965650 T8028 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1_shard1_0_replica1
   [junit4]   2> 3965653 T8039 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[118]} 0 18
   [junit4]   2> 3965653 T8028 oasc.SolrXMLCoresLocator.doPersist Persisted core descriptions to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/solr.xml
   [junit4]   2> 3965653 T8023 oasc.SolrXMLCoresLocator.doPersist Persisted core descriptions to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/solr.xml
   [junit4]   2> 3965654 T8023 oasc.CoreContainer.create Creating SolrCore 'collection1_shard1_1_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_1_replica1
   [junit4]   2> 3965655 T8023 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 3965654 T8028 oasc.CoreContainer.create Creating SolrCore 'collection1_shard1_0_replica1' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_0_replica1
   [junit4]   2> 3965657 T8028 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 3965657 T8023 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3965658 T8023 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 3965659 T8028 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3965660 T8028 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 3965661 T8023 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_1_replica1/'
   [junit4]   2> 3965664 T8028 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_0_replica1/'
   [junit4]   2> 3965700 T8006 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[105 (-1440564854542827520)]} 0 0
   [junit4]   2> 3965721 T8072 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[105 (-1440564854561701888)]} 0 0
   [junit4]   2> 3965725 T8044 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[105 (-1440564854561701888)]} 0 7
   [junit4]   2> 3965738 T8003 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[119 (1440564854582673408)]} 0 1
   [junit4]   2> 3965765 T8076 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[119 (1440564854598402048)]} 0 0
   [junit4]   2> 3965767 T8041 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[119 (1440564854598402048)]} 0 14
   [junit4]   2> 3965807 T8008 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[112 (-1440564854653976576)]} 0 1
   [junit4]   2> 3965821 T8074 C1816 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[112 (-1440564854665510912)]} 0 1
   [junit4]   2> 3965827 T8040 C1817 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[112 (-1440564854665510912)]} 0 10
   [junit4]   2> 3965837 T8004 C1818 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[120 (1440564854685433856)]} 0 2
   [junit4]   2> 3965886 T8028 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> ASYNC  NEW_CORE C1821 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1821_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3965890 T8077 C1821 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[120 (1440564854729474048)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1822 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1822_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3965901 T8043 C1822 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[120 (1440564854729474048)]} 0 23
   [junit4]   2> 3965900 T8023 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> ASYNC  NEW_CORE C1823 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1823_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3965944 T8005 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[121 (1440564854797631488)]} 0 2
   [junit4]   2> 3965964 T8075 C1821 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[121 (1440564854808117248)]} 0 1
   [junit4]   2> 3965968 T8039 C1822 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[121 (1440564854808117248)]} 0 15
   [junit4]   2> 3966019 T8007 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[122 (1440564854877323264)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1824 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1824_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3966035 T8053 C1824 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[122 (1440564854890954752)]} 0 0
   [junit4]   2> ASYNC  NEW_CORE C1825 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1825_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3966037 T8026 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[122 (1440564854890954752)]} 0 6
   [junit4]   2> 3966075 T8023 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3966079 T8003 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[117 (-1440564854940237824)]} 0 0
   [junit4]   2> 3966093 T8023 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3966101 T8072 C1821 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[117 (-1440564854957015040)]} 0 0
   [junit4]   2> 3966103 T8044 C1822 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[117 (-1440564854957015040)]} 0 9
   [junit4]   2> 3966105 T8024 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[117]} 0 13
   [junit4]   2> 3966115 T8028 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3966118 T8008 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[123 (1440564854980083712)]} 0 1
   [junit4]   2> 3966121 T8028 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3966140 T8028 oass.IndexSchema.readSchema [collection1_shard1_0_replica1] Schema name=test
   [junit4]   2> 3966151 T8056 C1824 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[123 (1440564855008395264)]} 0 1
   [junit4]   2> 3966155 T8025 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[123 (1440564855008395264)]} 0 12
   [junit4]   2> 3966158 T8042 C1822 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[123]} 0 28
   [junit4]   2> 3966173 T8023 oass.IndexSchema.readSchema [collection1_shard1_1_replica1] Schema name=test
   [junit4]   2> 3966199 T8004 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[109 (-1440564855066066944)]} 0 1
   [junit4]   2> 3966211 T8057 C1824 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[109 (-1440564855071309824)]} 0 0
   [junit4]   2> 3966214 T8026 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[109 (-1440564855071309824)]} 0 10
   [junit4]   2> 3966220 T8005 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[124 (1440564855087038464)]} 0 1
   [junit4]   2> 3966246 T8059 C1824 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[124 (1440564855109058560)]} 0 1
   [junit4]   2> 3966248 T8024 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[124 (1440564855109058560)]} 0 8
   [junit4]   2> 3966297 T8006 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[113 (-1440564855168827392)]} 0 1
   [junit4]   2> 3966312 T8055 C1824 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[113 (-1440564855178264576)]} 0 0
   [junit4]   2> 3966320 T8025 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[113 (-1440564855178264576)]} 0 14
   [junit4]   2> 3966337 T8007 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[125 (1440564855209721856)]} 0 1
   [junit4]   2> 3966355 T8053 C1824 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[125 (1440564855222304768)]} 0 1
   [junit4]   2> 3966357 T8026 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[125 (1440564855222304768)]} 0 10
   [junit4]   2> 3966400 T8003 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[122 (-1440564855276830720)]} 0 1
   [junit4]   2> 3966425 T8058 C1824 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[122 (-1440564855299899392)]} 0 0
   [junit4]   2> 3966427 T8024 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[122 (-1440564855299899392)]} 0 5
   [junit4]   2> 3966428 T8040 C1822 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[122]} 0 13
   [junit4]   2> 3966438 T8008 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[126 (1440564855315628032)]} 0 1
   [junit4]   2> 3966471 T8076 C1821 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[126 (1440564855347085312)]} 0 1
   [junit4]   2> 3966481 T8043 C1822 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[126 (1440564855347085312)]} 0 14
   [junit4]   2> 3966531 T8005 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[127 (1440564855407902720)]} 0 6
   [junit4]   2> 3966541 T8073 C1821 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[127 (1440564855418388480)]} 0 1
   [junit4]   2> 3966543 T8039 C1822 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[127 (1440564855418388480)]} 0 9
   [junit4]   2> 3966588 T8006 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[125 (-1440564855473963008)]} 0 0
   [junit4]   2> 3966613 T8056 C1824 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[125 (-1440564855488643072)]} 0 0
   [junit4]   2> 3966622 T8025 C1825 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[125 (-1440564855488643072)]} 0 20
   [junit4]   2> 3966632 T8044 C1822 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[125]} 0 37
   [junit4]   2> 3966651 T8007 C1823 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[128 (1440564855538974720)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1826 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1826_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3966684 T8077 C1826 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[128 (1440564855568334848)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1827 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1827_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3966686 T8041 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[128 (1440564855568334848)]} 0 8
   [junit4]   2> ASYNC  NEW_CORE C1828 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1828_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3966687 T8024 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[128]} 0 33
   [junit4]   2> ASYNC  NEW_CORE C1829 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1829_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3966730 T8003 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[129 (1440564855618666496)]} 0 4
   [junit4]   2> ASYNC  NEW_CORE C1830 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1830_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3966763 T8059 C1830 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[129 (1440564855638589440)]} 0 12
   [junit4]   2> 3966765 T8026 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[129 (1440564855638589440)]} 0 21
   [junit4]   2> 3966776 T8040 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[129]} 0 38
   [junit4]   2> 3966817 T8004 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[130 (1440564855714086912)]} 0 1
   [junit4]   2> 3966839 T8075 C1826 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[130 (1440564855731912704)]} 0 1
   [junit4]   2> 3966840 T8043 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[130 (1440564855731912704)]} 0 6
   [junit4]   2> 3966851 T8025 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[130]} 0 22
   [junit4]   2> 3966900 T8008 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[118 (-1440564855801118720)]} 0 1
   [junit4]   2> 3966908 T8055 C1830 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[118 (-1440564855805313024)]} 0 0
   [junit4]   2> 3966913 T8024 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[118 (-1440564855805313024)]} 0 9
   [junit4]   2> 3966925 T8006 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[131 (1440564855827333120)]} 0 1
   [junit4]   2> 3966955 T8072 C1826 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[131 (1440564855852498944)]} 0 1
   [junit4]   2> 3966957 T8039 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[131 (1440564855852498944)]} 0 8
   [junit4]   2> 3966961 T8026 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[131]} 0 16
   [junit4]   2> 3967013 T8007 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[132 (1440564855918559232)]} 0 2
   [junit4]   2> 3967038 T8053 C1830 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[132 (1440564855935336448)]} 0 6
   [junit4]   2> 3967040 T8025 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[132 (1440564855935336448)]} 0 13
   [junit4]   2> 3967047 T8042 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[132]} 0 28
   [junit4]   2> 3967102 T8003 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[133 (1440564856012931072)]} 0 1
   [junit4]   2> 3967124 T8057 C1830 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[133 (1440564856023416832)]} 0 1
   [junit4]   2> 3967125 T8026 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[133 (1440564856023416832)]} 0 14
   [junit4]   2> 3967171 T8004 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[134 (1440564856084234240)]} 0 1
   [junit4]   2> 3967187 T8076 C1826 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[134 (1440564856089477120)]} 0 5
   [junit4]   2> 3967191 T8041 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[134 (1440564856089477120)]} 0 16
   [junit4]   2> 3967237 T8005 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[135 (1440564856154488832)]} 0 1
   [junit4]   2> 3967260 T8056 C1830 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[135 (1440564856175460352)]} 0 0
   [junit4]   2> 3967262 T8024 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[135 (1440564856175460352)]} 0 6
   [junit4]   2> 3967269 T8040 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[135]} 0 21
   [junit4]   2> 3967309 T8008 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[136 (1440564856228937728)]} 0 1
   [junit4]   2> 3967336 T8073 C1826 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[136 (1440564856250957824)]} 0 1
   [junit4]   2> 3967340 T8043 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[136 (1440564856250957824)]} 0 12
   [junit4]   2> 3967350 T8025 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[136]} 0 28
   [junit4]   2> 3967396 T8007 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[133 (-1440564856320163840)]} 0 1
   [junit4]   2> 3967422 T8059 C1830 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[133 (-1440564856336941056)]} 0 4
   [junit4]   2> 3967427 T8026 C1828 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[133 (-1440564856336941056)]} 0 17
   [junit4]   2> 3967429 T8039 C1827 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[133]} 0 25
   [junit4]   2> 3967433 T8003 C1829 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[137 (1440564856358961152)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1831 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1831_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3967456 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[137 (1440564856377835520)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1832 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1832_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3967466 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[137 (1440564856377835520)]} 0 16
   [junit4]   2> ASYNC  NEW_CORE C1833 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1833_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3967527 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[138 (1440564856457527296)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1834 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1834_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3967558 T8055 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[138 (1440564856486887424)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1835 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1835_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3967564 T8025 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[138 (1440564856486887424)]} 0 11
   [junit4]   2> 3967574 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[138]} 0 33
   [junit4]   2> 3967622 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[139 (1440564856557142016)]} 0 1
   [junit4]   2> 3967657 T8074 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[139 (1440564856590696448)]} 0 1
   [junit4]   2> 3967661 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[139 (1440564856590696448)]} 0 8
   [junit4]   2> 3967666 T8026 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[139]} 0 18
   [junit4]   2> 3967712 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[107 (-1440564856651513856)]} 0 1
   [junit4]   2> 3967731 T8053 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[107 (-1440564856667242496)]} 0 0
   [junit4]   2> 3967733 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[107 (-1440564856667242496)]} 0 8
   [junit4]   2> 3967734 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[107]} 0 11
   [junit4]   2> 3967740 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[140 (1440564856680873984)]} 0 1
   [junit4]   2> 3967765 T8072 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[140 (1440564856698699776)]} 0 1
   [junit4]   2> 3967771 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[140 (1440564856698699776)]} 0 14
   [junit4]   2> 3967806 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[141 (1440564856750080000)]} 0 1
   [junit4]   2> 3967841 T8076 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[141 (1440564856781537280)]} 0 1
   [junit4]   2> 3967847 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[141 (1440564856781537280)]} 0 12
   [junit4]   2> 3967848 T8025 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[141]} 0 22
   [junit4]   2> 3967890 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[132 (-1440564856839208960)]} 0 0
   [junit4]   2> 3967895 T8023 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 3967897 T8057 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[132 (-1440564856843403264)]} 0 0
   [junit4]   2> 3967903 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[132 (-1440564856843403264)]} 0 9
   [junit4]   2> 3967906 T8023 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 3967911 T8023 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3967914 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[142 (1440564856863326208)]} 0 1
   [junit4]   2> 3967927 T8073 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[142 (1440564856872763392)]} 0 2
   [junit4]   2> 3967928 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[142 (1440564856872763392)]} 0 6
   [junit4]   2> 3967933 T8023 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3967971 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[140 (-1440564856924143616)]} 0 0
   [junit4]   2> 3967984 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[140 (-1440564856932532224)]} 0 1
   [junit4]   2> 3967986 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[140 (-1440564856932532224)]} 0 7
   [junit4]   2> 3967992 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[143 (1440564856945115136)]} 0 1
   [junit4]   2> 3968005 T8074 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[143 (1440564856954552320)]} 0 2
   [junit4]   2> 3968014 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[143 (1440564856954552320)]} 0 14
   [junit4]   2> 3968016 T8025 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[143]} 0 20
   [junit4]   2> 3968029 T8028 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 3968035 T8028 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 3968041 T8028 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3968056 T8028 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3968057 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[144 (1440564857012224000)]} 0 2
   [junit4]   2> 3968068 T8072 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[144 (1440564857019564032)]} 0 3
   [junit4]   2> 3968069 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[144 (1440564857019564032)]} 0 7
   [junit4]   2> 3968104 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[145 (1440564857062555648)]} 0 1
   [junit4]   2> 3968112 T8076 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[145 (1440564857066749952)]} 0 1
   [junit4]   2> 3968113 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[145 (1440564857066749952)]} 0 6
   [junit4]   2> 3968115 T8023 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3968116 T8023 oasc.SolrCore.<init> [collection1_shard1_1_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_1_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data/
   [junit4]   2> 3968116 T8023 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3968118 T8023 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data
   [junit4]   2> 3968119 T8023 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data/index/
   [junit4]   2> 3968119 T8023 oasc.SolrCore.initIndex WARN [collection1_shard1_1_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_1_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 3968123 T8023 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data/index
   [junit4]   2> 3968133 T8023 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_1,generation=1}
   [junit4]   2> 3968134 T8023 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3968145 T8023 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3968146 T8023 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3968147 T8023 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3968149 T8023 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3968149 T8023 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3968150 T8023 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3968150 T8023 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3968151 T8023 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3968152 T8023 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3968166 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[143 (-1440564857128615936)]} 0 1
   [junit4]   2> 3968180 T8073 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[143 (-1440564857140150272)]} 0 1
   [junit4]   2> 3968181 T8023 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3968186 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[143 (-1440564857140150272)]} 0 9
   [junit4]   2> 3968189 T8026 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[143]} 0 19
   [junit4]   2> 3968194 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[146 (1440564857156927488)]} 0 1
   [junit4]   2> 3968200 T8023 oass.SolrIndexSearcher.<init> Opening Searcher@6e3918db main
   [junit4]   2> 3968203 T8023 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3968204 T8023 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3968204 T8056 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[146 (1440564857164267520)]} 0 1
   [junit4]   2> 3968207 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[146 (1440564857164267520)]} 0 7
   [junit4]   2> 3968208 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[146]} 0 11
   [junit4]   2> 3968216 T8096 oasc.SolrCore.registerSearcher [collection1_shard1_1_replica1] Registered new searcher Searcher@6e3918db main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3968221 T8023 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 3968222 T8023 oasc.CoreContainer.registerCore registering core: collection1_shard1_1_replica1
   [junit4]   2> 3968222 T8023 oasc.ZkController.register Register replica - core:collection1_shard1_1_replica1 address:http://127.0.0.1:51255/qb collection:collection1 shard:shard1_1
   [junit4]   2> 3968223 T8023 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3968245 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[129 (-1440564857211453440)]} 0 0
   [junit4]   2> 3968254 T8028 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3968257 T8058 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[129 (-1440564857219842048)]} 0 0
   [junit4]   2> 3968259 T8025 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[129 (-1440564857219842048)]} 0 7
   [junit4]   2> 3968261 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[129]} 0 11
   [junit4]   2> 3968258 T8028 oasc.SolrCore.<init> [collection1_shard1_0_replica1] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_0_replica1/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data/
   [junit4]   2> 3968263 T8028 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3968261 T8023 oascc.SolrZkClient.makePath makePath: /collections/collection1/leader_elect/shard1_1/election
   [junit4]   2> 3968273 T8028 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data
   [junit4]   2> 3968274 T8028 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data/index/
   [junit4]   2> 3968275 T8028 oasc.SolrCore.initIndex WARN [collection1_shard1_0_replica1] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_0_replica1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 3968273 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[147 (1440564857240813568)]} 0 1
   [junit4]   2> 3968276 T8028 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data/index
   [junit4]   2> 3968284 T8075 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[147 (1440564857248153600)]} 0 1
   [junit4]   2> 3968286 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[147 (1440564857248153600)]} 0 6
   [junit4]   2> 3968291 T8028 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_1,generation=1}
   [junit4]   2> 3968293 T8028 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3968299 T8023 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1_1
   [junit4]   2> 3968306 T8028 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3968306 T8028 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3968307 T8028 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3968308 T8028 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3968309 T8028 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3968310 T8028 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3968310 T8028 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3968311 T8028 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3968312 T8028 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3968314 T8023 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 3968315 T8023 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 3968315 T8023 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/
   [junit4]   2> 3968316 T8023 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 3968317 T8023 oasc.SyncStrategy.syncToMe http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ has no replicas
   [junit4]   2> 3968318 T8023 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ shard1_1
   [junit4]   2> 3968318 T8023 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard1_1
   [junit4]   2> 3968330 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[148 (1440564857300582400)]} 0 0
   [junit4]   2> 3968337 T8028 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3968347 T8055 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[148 (1440564857314213888)]} 0 1
   [junit4]   2> 3968349 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[148 (1440564857314213888)]} 0 6
   [junit4]   2> 3968350 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[148]} 0 10
   [junit4]   2> 3968354 T8028 oass.SolrIndexSearcher.<init> Opening Searcher@5fd2b791 main
   [junit4]   2> 3968356 T8028 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3968356 T8028 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3968364 T8097 oasc.SolrCore.registerSearcher [collection1_shard1_0_replica1] Registered new searcher Searcher@5fd2b791 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3968368 T8028 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=null}
   [junit4]   2> 3968369 T8028 oasc.CoreContainer.registerCore registering core: collection1_shard1_0_replica1
   [junit4]   2> 3968371 T8028 oasc.ZkController.register Register replica - core:collection1_shard1_0_replica1 address:http://127.0.0.1:51255/qb collection:collection1 shard:shard1_0
   [junit4]   2> 3968372 T8028 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3968382 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[127 (-1440564857355108352)]} 0 0
   [junit4]   2> 3968405 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[127 (-1440564857376079872)]} 0 1
   [junit4]   2> 3968407 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[127 (-1440564857376079872)]} 0 5
   [junit4]   2> 3968408 T8025 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[127]} 0 15
   [junit4]   2> 3968410 T8028 oascc.SolrZkClient.makePath makePath: /collections/collection1/leader_elect/shard1_0/election
   [junit4]   2> 3968412 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[149 (1440564857386565632)]} 0 0
   [junit4]   2> 3968420 T8072 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[149 (1440564857390759936)]} 0 1
   [junit4]   2> 3968421 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[149 (1440564857390759936)]} 0 5
   [junit4]   2> 3968433 T8028 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1_0
   [junit4]   2> 3968439 T8028 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 3968440 T8028 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 3968440 T8028 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/
   [junit4]   2> 3968441 T8028 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 3968441 T8028 oasc.SyncStrategy.syncToMe http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ has no replicas
   [junit4]   2> 3968442 T8028 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ shard1_0
   [junit4]   2> 3968442 T8028 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard1_0
   [junit4]   2> 3968454 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[126 (-1440564857430605824)]} 0 0
   [junit4]   2> 3968464 T8076 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[126 (-1440564857438994432)]} 0 0
   [junit4]   2> 3968471 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[126 (-1440564857438994432)]} 0 10
   [junit4]   2> 3968472 T8026 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[126]} 0 14
   [junit4]   2> 3968477 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[150 (1440564857453674496)]} 0 1
   [junit4]   2> 3968490 T8053 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[150 (1440564857465208832)]} 0 1
   [junit4]   2> 3968491 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[150 (1440564857465208832)]} 0 5
   [junit4]   2> 3968492 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[150]} 0 8
   [junit4]   2> 3968526 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[142 (-1440564857506103296)]} 0 0
   [junit4]   2> 3968533 T8073 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[142 (-1440564857510297600)]} 0 1
   [junit4]   2> 3968534 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[142 (-1440564857510297600)]} 0 4
   [junit4]   2> 3968537 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[151 (1440564857517637632)]} 0 0
   [junit4]   2> 3968548 T8057 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[151 (1440564857524977664)]} 0 1
   [junit4]   2> 3968550 T8025 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[151 (1440564857524977664)]} 0 7
   [junit4]   2> 3968552 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[151]} 0 12
   [junit4]   2> 3968554 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3968589 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[152 (1440564857571115008)]} 0 1
   [junit4]   2> 3968601 T8075 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[152 (1440564857579503616)]} 0 1
   [junit4]   2> 3968607 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[152 (1440564857579503616)]} 0 11
   [junit4]   2> 3968614 T8026 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[152]} 0 21
   [junit4]   2> 3968621 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3968622 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3968622 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3968624 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3968624 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3968634 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3968661 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[153 (1440564857646612480)]} 0 1
   [junit4]   2> 3968672 T8074 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[153 (1440564857656049664)]} 0 1
   [junit4]   2> 3968674 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[153 (1440564857656049664)]} 0 6
   [junit4]   2> 3968675 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[153]} 0 10
   [junit4]   2> 3968683 T8023 oasc.ZkController.register We are http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ and leader is http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/
   [junit4]   2> 3968684 T8023 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 3968684 T8023 oasc.ZkController.publish publishing core=collection1_shard1_1_replica1 state=active
   [junit4]   2> 3968684 T8023 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3968685 T8028 oasc.ZkController.register We are http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ and leader is http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/
   [junit4]   2> 3968686 T8028 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 3968686 T8028 oasc.ZkController.publish publishing core=collection1_shard1_0_replica1 state=active
   [junit4]   2> 3968686 T8028 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3968689 T8023 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3968691 T8023 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={shard.range=c0000000-ffffffff&shard.state=construction&shard=shard1_1&name=collection1_shard1_1_replica1&action=CREATE&collection=collection1&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=4049 
   [junit4]   2> 3968692 T8028 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3968693 T8028 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={shard.range=80000000-bfffffff&shard.state=construction&shard=shard1_0&name=collection1_shard1_0_replica1&action=CREATE&collection=collection1&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=4055 
   [junit4]   2> 3968694 T8014 oasc.OverseerCollectionProcessor.splitShard Asking parent leader to wait for: collection1_shard1_0_replica1 to be alive on: 127.0.0.1:51255_qb
   [junit4]   2> 3968695 T8014 oasc.OverseerCollectionProcessor.splitShard Asking parent leader to wait for: collection1_shard1_1_replica1 to be alive on: 127.0.0.1:51255_qb
   [junit4]   2> 3968697 T8026 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node5, state: active, checkLive: true, onlyIfLeader: true
   [junit4]   2> 3968697 T8025 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node6, state: active, checkLive: true, onlyIfLeader: true
   [junit4]   2> 3968708 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[154 (1440564857696944128)]} 0 1
   [junit4]   2> 3968712 T8024 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 154
   [junit4]   2> ASYNC  NEW_CORE C1836 name=collection1_shard1_0_replica1 org.apache.solr.core.SolrCore@5c1f6bce url=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1 node=127.0.0.1:51255_qb C1836_STATE=coll:collection1 core:collection1_shard1_0_replica1 props:{state=down, core=collection1_shard1_0_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3968727 T8023 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[154 (1440564857700089856)]} 0 4
   [junit4]   2> 3968736 T8059 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[154 (1440564857700089856)]} 0 2
   [junit4]   2> 3968737 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[154 (1440564857700089856)]} 0 26
   [junit4]   2> 3968776 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[155 (1440564857767198720)]} 0 1
   [junit4]   2> 3968786 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[155 (1440564857774538752)]} 0 1
   [junit4]   2> 3968787 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[155 (1440564857774538752)]} 0 5
   [junit4]   2> 3968820 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[138 (-1440564857814384640)]} 0 0
   [junit4]   2> 3968824 T8028 C1835 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3968827 T8023 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[138 (-1440564857818578944)]} 0 0
   [junit4]   2> ASYNC  NEW_CORE C1837 name=collection1_shard1_1_replica1 org.apache.solr.core.SolrCore@54aa02ed url=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1 node=127.0.0.1:51255_qb C1837_STATE=coll:collection1 core:collection1_shard1_1_replica1 props:{state=down, core=collection1_shard1_1_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3968844 T8024 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[138 (-1440564857818578944)]} 0 3
   [junit4]   2> 3968854 T8058 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[138 (-1440564857818578944)]} 0 0
   [junit4]   2> 3968855 T8028 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[138 (-1440564857818578944)]} 0 32
   [junit4]   2> 3968864 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[156 (1440564857859473408)]} 0 1
   [junit4]   2> 3968872 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 156
   [junit4]   2> 3968877 T8024 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[156 (1440564857868910592)]} 0 1
   [junit4]   2> 3968882 T8055 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[156 (1440564857868910592)]} 0 1
   [junit4]   2> 3968883 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[156 (1440564857868910592)]} 0 12
   [junit4]   2> 3968884 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[156]} 0 15
   [junit4]   2> 3968918 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[153 (-1440564857916096512)]} 0 1
   [junit4]   2> 3968927 T8072 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[153 (-1440564857923436544)]} 0 1
   [junit4]   2> 3968930 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[153 (-1440564857923436544)]} 0 6
   [junit4]   2> 3968931 T8028 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[153]} 0 10
   [junit4]   2> 3968935 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[157 (1440564857934970880)]} 0 0
   [junit4]   2> 3968943 T8024 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 157
   [junit4]   2> 3968947 T8023 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[157 (1440564857942310912)]} 0 1
   [junit4]   2> 3968952 T8053 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[157 (1440564857942310912)]} 0 1
   [junit4]   2> 3968953 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[157 (1440564857942310912)]} 0 11
   [junit4]   2> 3968954 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[157]} 0 14
   [junit4]   2> 3968996 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[158 (1440564857998934016)]} 0 1
   [junit4]   2> 3969008 T8073 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[158 (1440564858004176896)]} 0 1
   [junit4]   2> 3969010 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[158 (1440564858004176896)]} 0 9
   [junit4]   2> 3969044 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[159 (1440564858048217088)]} 0 1
   [junit4]   2> 3969062 T8075 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[159 (1440564858060800000)]} 0 1
   [junit4]   2> 3969063 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[159 (1440564858060800000)]} 0 8
   [junit4]   2> 3969065 T8028 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[159]} 0 15
   [junit4]   2> 3969098 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[160 (1440564858105888768)]} 0 1
   [junit4]   2> 3969106 T8024 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 160
   [junit4]   2> 3969114 T8028 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[160 (1440564858112180224)]} 0 1
   [junit4]   2> 3969118 T8057 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[160 (1440564858112180224)]} 0 0
   [junit4]   2> 3969120 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[160 (1440564858112180224)]} 0 16
   [junit4]   2> 3969121 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[160]} 0 19
   [junit4]   2> 3969159 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[161 (1440564858168803328)]} 0 1
   [junit4]   2> 3969163 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 161
   [junit4]   2> 3969172 T8028 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[161 (1440564858174046208)]} 0 4
   [junit4]   2> 3969179 T8056 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[161 (1440564858174046208)]} 0 1
   [junit4]   2> 3969180 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[161 (1440564858174046208)]} 0 18
   [junit4]   2> 3969215 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[162 (1440564858227523584)]} 0 1
   [junit4]   2> 3969226 T8074 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[162 (1440564858235912192)]} 0 0
   [junit4]   2> 3969228 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[162 (1440564858235912192)]} 0 6
   [junit4]   2> 3969230 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[162]} 0 11
   [junit4]   2> 3969263 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[163 (1440564858278903808)]} 0 0
   [junit4]   2> 3969274 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[163 (1440564858286243840)]} 0 1
   [junit4]   2> 3969276 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[163 (1440564858286243840)]} 0 7
   [junit4]   2> 3969277 T8028 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[163]} 0 10
   [junit4]   2> 3969315 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[161 (-1440564858332381184)]} 0 1
   [junit4]   2> 3969327 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3969330 T8024 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[161 (-1440564858346012672)]} 0 0
   [junit4]   2> 3969336 T8028 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[161 (-1440564858346012672)]} 0 1
   [junit4]   2> 3969340 T8058 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[161 (-1440564858346012672)]} 0 0
   [junit4]   2> 3969342 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[161 (-1440564858346012672)]} 0 15
   [junit4]   2> 3969344 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[161]} 0 23
   [junit4]   2> 3969348 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[164 (1440564858368032768)]} 0 0
   [junit4]   2> 3969358 T8072 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[164 (1440564858372227072)]} 0 1
   [junit4]   2> 3969369 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[164 (1440564858372227072)]} 0 17
   [junit4]   2> 3969416 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[115 (-1440564858438287360)]} 0 1
   [junit4]   2> 3969436 T8073 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[115 (-1440564858448773120)]} 0 0
   [junit4]   2> 3969438 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[115 (-1440564858448773120)]} 0 13
   [junit4]   2> 3969455 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[165 (1440564858479181824)]} 0 1
   [junit4]   2> 3969461 T8075 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[165 (1440564858483376128)]} 0 0
   [junit4]   2> 3969463 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[165 (1440564858483376128)]} 0 5
   [junit4]   2> 3969495 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[162 (-1440564858522173440)]} 0 0
   [junit4]   2> 3969504 T8074 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[162 (-1440564858527416320)]} 0 1
   [junit4]   2> 3969505 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[162 (-1440564858527416320)]} 0 5
   [junit4]   2> 3969507 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[162]} 0 9
   [junit4]   2> 3969516 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[166 (1440564858543144960)]} 0 1
   [junit4]   2> 3969528 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 166
   [junit4]   2> 3969532 T8028 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[166 (1440564858556776448)]} 0 1
   [junit4]   2> 3969537 T8055 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[166 (1440564858556776448)]} 0 1
   [junit4]   2> 3969538 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[166 (1440564858556776448)]} 0 11
   [junit4]   2> 3969540 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[166]} 0 17
   [junit4]   2> 3969579 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[155 (-1440564858609205248)]} 0 1
   [junit4]   2> 3969586 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[155 (-1440564858614448128)]} 0 1
   [junit4]   2> 3969587 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[155 (-1440564858614448128)]} 0 4
   [junit4]   2> 3969592 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[167 (1440564858622836736)]} 0 0
   [junit4]   2> 3969598 T8024 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 167
   [junit4]   2> 3969602 T8028 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[167 (1440564858630176768)]} 0 0
   [junit4]   2> 3969613 T8053 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[167 (1440564858630176768)]} 0 1
   [junit4]   2> 3969615 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[167 (1440564858630176768)]} 0 17
   [junit4]   2> 3969616 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[167]} 0 21
   [junit4]   2> 3969650 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[168 (1440564858683654144)]} 0 2
   [junit4]   2> 3969664 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 168
   [junit4]   2> 3969668 T8028 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[168 (1440564858699382784)]} 0 1
   [junit4]   2> 3969683 T8059 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[168 (1440564858699382784)]} 0 1
   [junit4]   2> 3969693 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[168 (1440564858699382784)]} 0 29
   [junit4]   2> 3969732 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[169 (1440564858769637376)]} 0 1
   [junit4]   2> 3969739 T8024 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 169
   [junit4]   2> 3969743 T8028 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[169 (1440564858778025984)]} 0 0
   [junit4]   2> 3969757 T8056 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[169 (1440564858778025984)]} 0 1
   [junit4]   2> 3969758 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[169 (1440564858778025984)]} 0 20
   [junit4]   2> 3969767 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[169]} 0 31
   [junit4]   2> 3969811 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[165 (-1440564858853523456)]} 0 0
   [junit4]   2> 3969818 T8076 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[165 (-1440564858857717760)]} 0 0
   [junit4]   2> 3969820 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[165 (-1440564858857717760)]} 0 5
   [junit4]   2> 3969824 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[170 (1440564858866106368)]} 0 1
   [junit4]   2> 3969831 T8073 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[170 (1440564858871349248)]} 0 0
   [junit4]   2> 3969833 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[170 (1440564858871349248)]} 0 5
   [junit4]   2> 3969866 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[171 (1440564858911195136)]} 0 1
   [junit4]   2> 3969871 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 171
   [junit4]   2> 3969875 T8028 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[171 (1440564858916438016)]} 0 1
   [junit4]   2> 3969880 T8058 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[171 (1440564858916438016)]} 0 1
   [junit4]   2> 3969882 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[171 (1440564858916438016)]} 0 12
   [junit4]   2> 3969915 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[135 (-1440564858962575360)]} 0 0
   [junit4]   2> 3969934 T8024 C1835 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3969937 T8023 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[135 (-1440564858981449728)]} 0 0
   [junit4]   2> 3969941 T8028 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[135 (-1440564858981449728)]} 0 0
   [junit4]   2> 3969945 T8055 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[135 (-1440564858981449728)]} 0 1
   [junit4]   2> 3969946 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[135 (-1440564858981449728)]} 0 13
   [junit4]   2> 3969947 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[135]} 0 22
   [junit4]   2> 3969952 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[172 (1440564859000324096)]} 0 1
   [junit4]   2> 3969959 T8075 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[172 (1440564859005566976)]} 0 1
   [junit4]   2> 3969960 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[172 (1440564859005566976)]} 0 5
   [junit4]   2> 3969999 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[173 (1440564859050655744)]} 0 1
   [junit4]   2> 3970006 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 173
   [junit4]   2> 3970010 T8028 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[173 (1440564859057995776)]} 0 1
   [junit4]   2> 3970014 T8053 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[173 (1440564859057995776)]} 0 1
   [junit4]   2> 3970015 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[173 (1440564859057995776)]} 0 9
   [junit4]   2> 3970016 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[173]} 0 13
   [junit4]   2> 3970054 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[145 (-1440564859108327424)]} 0 0
   [junit4]   2> 3970070 T8074 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[145 (-1440564859121958912)]} 0 0
   [junit4]   2> 3970072 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[145 (-1440564859121958912)]} 0 6
   [junit4]   2> 3970073 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[145]} 0 11
   [junit4]   2> 3970077 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[174 (1440564859132444672)]} 0 1
   [junit4]   2> 3970085 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[174 (1440564859136638976)]} 0 1
   [junit4]   2> 3970086 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[174 (1440564859136638976)]} 0 5
   [junit4]   2> 3970124 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[175 (1440564859180679168)]} 0 1
   [junit4]   2> 3970134 T8028 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 175
   [junit4]   2> 3970138 T8023 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[175 (1440564859192213504)]} 0 1
   [junit4]   2> 3970142 T8059 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[175 (1440564859192213504)]} 0 1
   [junit4]   2> 3970143 T8028 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[175 (1440564859192213504)]} 0 10
   [junit4]   2> 3970144 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[175]} 0 14
   [junit4]   2> 3970187 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[116 (-1440564859246739456)]} 0 2
   [junit4]   2> 3970194 T8072 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[116 (-1440564859250933760)]} 0 0
   [junit4]   2> 3970195 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3970196 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[116 (-1440564859250933760)]} 0 6
   [junit4]   2> 3970199 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":null,
   [junit4]   2> 	  "core":"collection1_shard1_1_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3970201 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[176 (1440564859262468096)]} 0 1
   [junit4]   2> 3970207 T8024 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 176
   [junit4]   2> 3970211 T8023 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[176 (1440564859267710976)]} 0 0
   [junit4]   2> 3970226 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_0",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":null,
   [junit4]   2> 	  "core":"collection1_shard1_0_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3970230 T8056 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[176 (1440564859267710976)]} 0 1
   [junit4]   2> 3970232 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[176 (1440564859267710976)]} 0 26
   [junit4]   2> 3970244 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3970245 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3970246 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3970247 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3970247 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3970247 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3970273 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[149 (-1440564859337965568)]} 0 0
   [junit4]   2> 3970284 T8076 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[149 (-1440564859343208448)]} 0 1
   [junit4]   2> 3970285 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[149 (-1440564859343208448)]} 0 7
   [junit4]   2> 3970291 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[177 (1440564859356839936)]} 0 0
   [junit4]   2> 3970305 T8075 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[177 (1440564859367325696)]} 0 1
   [junit4]   2> 3970307 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[177 (1440564859367325696)]} 0 6
   [junit4]   2> 3970309 T8028 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[177]} 0 11
   [junit4]   2> 3970347 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[178 (1440564859412414464)]} 0 5
   [junit4]   2> 3970362 T8074 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[178 (1440564859427094528)]} 0 1
   [junit4]   2> 3970363 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[178 (1440564859427094528)]} 0 6
   [junit4]   2> 3970395 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[128 (-1440564859465891840)]} 0 0
   [junit4]   2> 3970402 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[128 (-1440564859471134720)]} 0 0
   [junit4]   2> 3970404 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[128 (-1440564859471134720)]} 0 4
   [junit4]   2> 3970408 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[179 (1440564859478474752)]} 0 1
   [junit4]   2> 3970421 T8072 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[179 (1440564859485814784)]} 0 1
   [junit4]   2> 3970423 T8043 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[179 (1440564859485814784)]} 0 11
   [junit4]   2> 3970455 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[164 (-1440564859528806400)]} 0 0
   [junit4]   2> 3970461 T8073 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[164 (-1440564859533000704)]} 0 0
   [junit4]   2> 3970463 T8039 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[164 (-1440564859533000704)]} 0 4
   [junit4]   2> 3970467 T8007 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[180 (1440564859540340736)]} 0 1
   [junit4]   2> 3970472 T8076 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[180 (1440564859543486464)]} 0 0
   [junit4]   2> 3970474 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[180 (1440564859543486464)]} 0 5
   [junit4]   2> 3970510 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[181 (1440564859585429504)]} 0 1
   [junit4]   2> 3970520 T8075 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[181 (1440564859592769536)]} 0 1
   [junit4]   2> 3970522 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[181 (1440564859592769536)]} 0 6
   [junit4]   2> 3970523 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[181]} 0 9
   [junit4]   2> 3970557 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[182 (1440564859635761152)]} 0 1
   [junit4]   2> 3970561 T8028 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 182
   [junit4]   2>  C1837_STATE=coll:collection1 core:collection1_shard1_1_replica1 props:{state=active, core=collection1_shard1_1_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3970565 T8024 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[182 (1440564859639955456)]} 0 0
   [junit4]   2> 3970569 T8058 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[182 (1440564859639955456)]} 0 0
   [junit4]   2> 3970570 T8028 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[182 (1440564859639955456)]} 0 9
   [junit4]   2> 3970604 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[183 (1440564859683995648)]} 0 1
   [junit4]   2> 3970617 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 183
   [junit4]   2>  C1836_STATE=coll:collection1 core:collection1_shard1_0_replica1 props:{state=active, core=collection1_shard1_0_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3970620 T8024 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[183 (1440564859697627136)]} 0 0
   [junit4]   2> 3970628 T8057 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[183 (1440564859697627136)]} 0 1
   [junit4]   2> 3970629 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[183 (1440564859697627136)]} 0 13
   [junit4]   2> 3970630 T8042 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[183]} 0 17
   [junit4]   2> 3970663 T8008 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[121 (-1440564859746910208)]} 0 1
   [junit4]   2> 3970673 T8074 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[121 (-1440564859754250240)]} 0 1
   [junit4]   2> 3970674 T8041 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[121 (-1440564859754250240)]} 0 4
   [junit4]   2> 3970675 T8028 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[121]} 0 9
   [junit4]   2> 3970681 T8003 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[184 (1440564859762638848)]} 0 3
   [junit4]   2> 3970689 T8024 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 184
   [junit4]   2> 3970697 T8023 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[184 (1440564859774173184)]} 0 1
   [junit4]   2> 3970700 T8026 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node5, state: active, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 3970701 T8026 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node5&state=active&nodeName=127.0.0.1:51255_qb&action=PREPRECOVERY&checkLive=true&core=collection1_shard1_0_replica1&wt=javabin&qt=/admin/cores&onlyIfLeader=true&version=2} status=0 QTime=2004 
   [junit4]   2> 3970700 T8025 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node6, state: active, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 3970702 T8025 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node6&state=active&nodeName=127.0.0.1:51255_qb&action=PREPRECOVERY&checkLive=true&core=collection1_shard1_1_replica1&wt=javabin&qt=/admin/cores&onlyIfLeader=true&version=2} status=0 QTime=2005 
   [junit4]   2> 3970702 T8053 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[184 (1440564859774173184)]} 0 1
   [junit4]   2> 3970704 T8014 oasc.OverseerCollectionProcessor.splitShard Successfully created all sub-shards for collection collection1 parent shard: shard1 on: core_node1:{"state":"active","core":"collection1","node_name":"127.0.0.1:51255_qb","base_url":"http://127.0.0.1:51255/qb","leader":"true"}
   [junit4]   2> 3970704 T8014 oasc.OverseerCollectionProcessor.splitShard Splitting shard core_node1 as part of slice shard1 of collection collection1 on core_node1:{"state":"active","core":"collection1","node_name":"127.0.0.1:51255_qb","base_url":"http://127.0.0.1:51255/qb","leader":"true"}
   [junit4]   2> 3970706 T8024 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[184 (1440564859774173184)]} 0 17
   [junit4]   2> 3970708 T8028 oasha.CoreAdminHandler.handleSplitAction Invoked split action for core: collection1
   [junit4]   2> 3970708 T8028 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3970713 T8028 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=3
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_2,generation=2}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_3,generation=3}
   [junit4]   2> 3970715 T8028 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3970717 T8028 oass.SolrIndexSearcher.<init> Opening Searcher@3f88774b main
   [junit4]   2> 3970717 T8028 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3970718 T8034 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@3f88774b main{StandardDirectoryReader(segments_3:5:nrt _0(5.0):C59 _1(5.0):C39/18)}
   [junit4]   2> 3970719 T8028 oasu.SolrIndexSplitter.split SolrIndexSplitter: partitions=2 segments=2
   [junit4]   2> 3970720 T8028 oasu.SolrIndexSplitter.split SolrIndexSplitter: partition #0 range=80000000-bfffffff
   [junit4]   2> 3970722 T8028 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_1,generation=1}
   [junit4]   2> 3970723 T8028 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3970726 T8028 oasu.SolrIndexSplitter.split SolrIndexSplitter: partition #1 range=c0000000-ffffffff
   [junit4]   2> 3970733 T8028 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_1,generation=1}
   [junit4]   2> 3970735 T8028 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3970740 T8006 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[185 (1440564859827650560)]} 0 1
   [junit4]   2> 3970741 T8028 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={targetCore=collection1_shard1_0_replica1&targetCore=collection1_shard1_1_replica1&action=SPLIT&core=collection1&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=33 
   [junit4]   2> 3970744 T8014 oasc.OverseerCollectionProcessor.splitShard Index on shard: 127.0.0.1:51255_qb split into two successfully
   [junit4]   2> 3970744 T8014 oasc.OverseerCollectionProcessor.splitShard Applying buffered updates on : collection1_shard1_0_replica1
   [junit4]   2> 3970744 T8014 oasc.OverseerCollectionProcessor.splitShard Applying buffered updates on : collection1_shard1_1_replica1
   [junit4]   2> 3970751 T8098 C1837 P51255 oasu.UpdateLog$LogReplayer.doReplay WARN Starting log replay tlog{file=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_1_replica1/data/tlog/tlog.0000000000000000000 refcount=2} active=true starting pos=0
   [junit4]   2> 3970751 T8099 C1836 P51255 oasu.UpdateLog$LogReplayer.doReplay WARN Starting log replay tlog{file=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033/collection1_shard1_0_replica1/data/tlog/tlog.0000000000000000000 refcount=2} active=true starting pos=0
   [junit4]   2> 3970762 T8023 C1835 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 185
   [junit4]   2> 3970765 T8098 C1837 P51255 oasu.DirectUpdateHandler2.commit start commit{flags=2,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3970765 T8099 C1836 P51255 oasu.DirectUpdateHandler2.commit start commit{flags=2,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3970770 T8099 C1836 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_2,generation=2}
   [junit4]   2> 3970774 T8099 C1836 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 2
   [junit4]   2> 3970779 T8099 C1836 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@58fd8d93 main
   [junit4]   2> 3970781 T8098 C1837 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_2,generation=2}
   [junit4]   2> 3970781 T8098 C1837 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 2
   [junit4]   2> 3970783 T8098 C1837 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@406ad619 main
   [junit4]   2> 3970784 T8099 C1836 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3970784 T8097 oasc.SolrCore.registerSearcher [collection1_shard1_0_replica1] Registered new searcher Searcher@58fd8d93 main{StandardDirectoryReader(segments_2:6:nrt _0(5.0):C26/8 _1(5.0):C8)}
   [junit4]   2> 3970785 T8096 oasc.SolrCore.registerSearcher [collection1_shard1_1_replica1] Registered new searcher Searcher@406ad619 main{StandardDirectoryReader(segments_2:6:nrt _0(5.0):C54/7 _1(5.0):C8/1)}
   [junit4]   2> 3970786 T8099 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] {add=[154 (1440564857700089856), 156 (1440564857868910592), 160 (1440564858112180224), 166 (1440564858556776448), 171 (1440564858916438016), 175 (1440564859192213504), 183 (1440564859697627136), 184 (1440564859774173184)],delete=[138 (-1440564857818578944), 161 (-1440564858346012672), 135 (-1440564858981449728)]} 0 35
   [junit4]   2> 3970787 T8099 C1836 P51255 oasu.UpdateLog$LogReplayer.run WARN Log replay finished. recoveryInfo=RecoveryInfo{adds=8 deletes=3 deleteByQuery=0 errors=0 positionOfStart=0}
   [junit4]   2> 3970787 T8025 oasc.ZkController.publish publishing core=collection1_shard1_0_replica1 state=active
   [junit4]   2> 3970788 T8025 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3970787 T8098 C1837 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3970791 T8098 C1837 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] {delete=[138 (-1440564857818578944), 161 (-1440564858346012672), 135 (-1440564858981449728)],add=[157 (1440564857942310912), 161 (1440564858174046208), 167 (1440564858630176768), 168 (1440564858699382784), 169 (1440564858778025984), 173 (1440564859057995776), 176 (1440564859267710976), 182 (1440564859639955456)]} 0 40
   [junit4]   2> 3970791 T8098 C1837 P51255 oasu.UpdateLog$LogReplayer.run WARN Log replay finished. recoveryInfo=RecoveryInfo{adds=8 deletes=3 deleteByQuery=0 errors=0 positionOfStart=0}
   [junit4]   2> 3970790 T8025 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={name=collection1_shard1_0_replica1&action=REQUESTAPPLYUPDATES&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=40 
   [junit4]   2> 3970790 T8024 C1836 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[185 (1440564859833942016)]} 0 24
   [junit4]   2> 3970792 T8026 oasc.ZkController.publish publishing core=collection1_shard1_1_replica1 state=active
   [junit4]   2> 3970794 T8026 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3970797 T8059 C1834 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[185 (1440564859833942016)]} 0 1
   [junit4]   2> 3970798 T8026 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={name=collection1_shard1_1_replica1&action=REQUESTAPPLYUPDATES&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=48 
   [junit4]   2> 3970798 T8023 C1835 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[185 (1440564859833942016)]} 0 52
   [junit4]   2> 3970800 T8014 oasc.OverseerCollectionProcessor.splitShard Successfully applied buffered updates on : [collection1_shard1_0_replica1, collection1_shard1_1_replica1]
   [junit4]   2> 3970801 T8014 oasc.OverseerCollectionProcessor.splitShard Creating replica shard collection1_shard1_0_replica2 as part of slice shard1_0 of collection collection1 on 127.0.0.1:51258_qb
   [junit4]   2> 3970805 T8039 oasc.ZkController.publish publishing core=collection1_shard1_0_replica2 state=down
   [junit4]   2> 3970805 T8039 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3970808 T8039 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 3970834 T8005 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[186 (1440564859926216704)]} 0 0
   [junit4]   2> 3970845 T8077 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[186 (1440564859930411008)]} 0 1
   [junit4]   2> 3970847 T8040 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[186 (1440564859930411008)]} 0 9
   [junit4]   2> 3970888 T8004 C1833 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[187 (1440564859977596928)]} 0 7
   [junit4]   2> 3970901 T8073 C1831 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[187 (1440564859992276992)]} 0 1
   [junit4]   2> 3970902 T8044 C1832 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[187 (1440564859992276992)]} 0 5
   [junit4]   2> ASYNC  NEW_CORE C1838 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1838_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3970919 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[187]} 0 26
   [junit4]   2> ASYNC  NEW_CORE C1839 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1839_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3970952 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[171 (-1440564860048900096)]} 0 1
   [junit4]   2> 3970963 T8024 C1838 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> ASYNC  NEW_CORE C1840 name=collection1_shard1_0_replica1 org.apache.solr.core.SolrCore@5c1f6bce url=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1 node=127.0.0.1:51255_qb C1840_STATE=coll:collection1 core:collection1_shard1_0_replica1 props:{state=active, core=collection1_shard1_0_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3970968 T8026 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[171 (-1440564860061483008)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1841 name=collection1_shard1_1_replica1 org.apache.solr.core.SolrCore@54aa02ed url=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1 node=127.0.0.1:51255_qb C1841_STATE=coll:collection1 core:collection1_shard1_1_replica1 props:{state=active, core=collection1_shard1_1_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3970978 T8023 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[171 (-1440564860061483008)]} 0 2
   [junit4]   2> ASYNC  NEW_CORE C1842 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1842_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3970982 T8056 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[171 (-1440564860061483008)]} 0 0
   [junit4]   2> 3970984 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[171 (-1440564860061483008)]} 0 20
   [junit4]   2> ASYNC  NEW_CORE C1843 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1843_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3970985 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[171]} 0 25
   [junit4]   2> 3970990 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[188 (1440564860088745984)]} 0 1
   [junit4]   2> 3970996 T8028 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 188
   [junit4]   2> 3971000 T8025 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[188 (1440564860095037440)]} 0 1
   [junit4]   2> 3971007 T8058 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[188 (1440564860095037440)]} 0 1
   [junit4]   2> 3971008 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[188 (1440564860095037440)]} 0 13
   [junit4]   2> 3971009 T8041 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[188]} 0 16
   [junit4]   2> 3971042 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[184 (-1440564860144320512)]} 0 0
   [junit4]   2> 3971047 T8026 C1838 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3971051 T8023 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[184 (-1440564860148514816)]} 0 1
   [junit4]   2> 3971054 T8024 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[184 (-1440564860148514816)]} 0 0
   [junit4]   2> 3971057 T8057 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[184 (-1440564860148514816)]} 0 0
   [junit4]   2> 3971059 T8026 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[184 (-1440564860148514816)]} 0 13
   [junit4]   2> 3971063 T8006 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[189 (1440564860165292032)]} 0 1
   [junit4]   2> 3971069 T8025 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 189
   [junit4]   2> 3971073 T8028 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[189 (1440564860172632064)]} 0 1
   [junit4]   2> 3971078 T8053 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[189 (1440564860172632064)]} 0 1
   [junit4]   2> 3971079 T8025 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[189 (1440564860172632064)]} 0 11
   [junit4]   2> 3971082 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[189]} 0 16
   [junit4]   2> 3971116 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[187 (-1440564860221915136)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1844 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1844_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3971124 T8076 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[187 (-1440564860227158016)]} 0 1
   [junit4]   2> 3971125 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[187 (-1440564860227158016)]} 0 4
   [junit4]   2> 3971130 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[190 (1440564860235546624)]} 0 1
   [junit4]   2> 3971138 T8023 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 190
   [junit4]   2> 3971150 T8024 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[190 (1440564860244983808)]} 0 1
   [junit4]   2> 3971156 T8059 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[190 (1440564860244983808)]} 0 1
   [junit4]   2> 3971158 T8023 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[190 (1440564860244983808)]} 0 20
   [junit4]   2> 3971160 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[190]} 0 25
   [junit4]   2> 3971209 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[191 (1440564860318384128)]} 0 1
   [junit4]   2> 3971229 T8075 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[191 (1440564860323627008)]} 0 5
   [junit4]   2> 3971232 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[191 (1440564860323627008)]} 0 19
   [junit4]   2> 3971268 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[192 (1440564860380250112)]} 0 1
   [junit4]   2> 3971272 T8026 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 192
   [junit4]   2> 3971281 T8028 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[192 (1440564860385492992)]} 0 3
   [junit4]   2> 3971286 T8055 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[192 (1440564860385492992)]} 0 1
   [junit4]   2> 3971287 T8026 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[192 (1440564860385492992)]} 0 16
   [junit4]   2> 3971322 T8006 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[193 (1440564860436873216)]} 0 1
   [junit4]   2> 3971330 T8025 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 193
   [junit4]   2> 3971336 T8023 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[193 (1440564860446310400)]} 0 1
   [junit4]   2> 3971343 T8056 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[193 (1440564860446310400)]} 0 2
   [junit4]   2> 3971344 T8025 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[193 (1440564860446310400)]} 0 15
   [junit4]   2> 3971346 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[193]} 0 20
   [junit4]   2> 3971383 T8004 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[139 (-1440564860500836352)]} 0 1
   [junit4]   2> 3971392 T8074 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[139 (-1440564860508176384)]} 0 0
   [junit4]   2> 3971394 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[139 (-1440564860508176384)]} 0 5
   [junit4]   2> 3971395 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[139]} 0 9
   [junit4]   2> 3971399 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[194 (1440564860517613568)]} 0 1
   [junit4]   2> 3971406 T8077 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[194 (1440564860521807872)]} 0 1
   [junit4]   2> 3971407 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[194 (1440564860521807872)]} 0 5
   [junit4]   2> 3971440 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[179 (-1440564860561653760)]} 0 0
   [junit4]   2> 3971454 T8073 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[179 (-1440564860570042368)]} 0 1
   [junit4]   2> 3971455 T8041 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[179 (-1440564860570042368)]} 0 8
   [junit4]   2> 3971457 T8026 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[179]} 0 13
   [junit4]   2> 3971461 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[195 (1440564860582625280)]} 0 1
   [junit4]   2> 3971473 T8076 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[195 (1440564860593111040)]} 0 0
   [junit4]   2> 3971474 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[195 (1440564860593111040)]} 0 4
   [junit4]   2> 3971508 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[154 (-1440564860631908352)]} 0 1
   [junit4]   2> 3971520 T8024 C1838 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3971528 T8023 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[154 (-1440564860645539840)]} 0 0
   [junit4]   2> 3971535 T8025 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[154 (-1440564860645539840)]} 0 1
   [junit4]   2> 3971539 T8058 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[154 (-1440564860645539840)]} 0 0
   [junit4]   2> 3971540 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[154 (-1440564860645539840)]} 0 20
   [junit4]   2> 3971542 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[154]} 0 28
   [junit4]   2> 3971546 T8006 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[196 (1440564860671754240)]} 0 1
   [junit4]   2> 3971558 T8075 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[196 (1440564860681191424)]} 0 1
   [junit4]   2> 3971559 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[196 (1440564860681191424)]} 0 5
   [junit4]   2> 3971561 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[196]} 0 10
   [junit4]   2> 3971602 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[173 (-1440564860731523072)]} 0 1
   [junit4]   2> 3971608 T8026 C1838 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3971612 T8023 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[173 (-1440564860737814528)]} 0 1
   [junit4]   2> 3971621 T8025 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[173 (-1440564860737814528)]} 0 0
   [junit4]   2> 3971641 T8053 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[173 (-1440564860737814528)]} 0 1
   [junit4]   2> 3971642 T8026 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[173 (-1440564860737814528)]} 0 34
   [junit4]   2> 3971643 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[173]} 0 38
   [junit4]   2> 3971648 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[197 (1440564860778708992)]} 0 1
   [junit4]   2> 3971654 T8024 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 197
   [junit4]   2> 3971662 T8028 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[197 (1440564860786049024)]} 0 1
   [junit4]   2> 3971666 T8059 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[197 (1440564860786049024)]} 0 1
   [junit4]   2> 3971667 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[197 (1440564860786049024)]} 0 14
   [junit4]   2> 3971668 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[197]} 0 17
   [junit4]   2> 3971702 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[198 (1440564860835332096)]} 0 1
   [junit4]   2> 3971709 T8072 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[198 (1440564860840574976)]} 0 1
   [junit4]   2> 3971714 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[198 (1440564860840574976)]} 0 9
   [junit4]   2> 3971754 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[152 (-1440564860890906624)]} 0 0
   [junit4]   2> 3971761 T8074 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[152 (-1440564860895100928)]} 0 0
   [junit4]   2> 3971764 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[152 (-1440564860895100928)]} 0 6
   [junit4]   2> 3971768 T8004 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[199 (1440564860904538112)]} 0 1
   [junit4]   2> 3971775 T8077 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[199 (1440564860909780992)]} 0 1
   [junit4]   2> 3971777 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[199 (1440564860909780992)]} 0 6
   [junit4]   2> 3971782 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3971790 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_0",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":null,
   [junit4]   2> 	  "core":"collection1_shard1_0_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3971806 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":null,
   [junit4]   2> 	  "core":"collection1_shard1_1_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3971813 T8006 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[198 (-1440564860952772608)]} 0 0
   [junit4]   2> 3971821 T8073 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[198 (-1440564860956966912)]} 0 0
   [junit4]   2> 3971822 T8041 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[198 (-1440564860956966912)]} 0 5
   [junit4]   2> 3971826 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[200 (1440564860966404096)]} 0 0
   [junit4]   2> 3971827 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_0",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1_shard1_0_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 3971833 T8075 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[200 (1440564860970598400)]} 0 1
   [junit4]   2> 3971834 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[200 (1440564860970598400)]} 0 5
   [junit4]   2> 3971842 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3971843 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3971843 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3971843 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3971844 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3971844 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3971872 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[201 (1440564861013590016)]} 0 1
   [junit4]   2> 3971877 T8023 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 201
   [junit4]   2> 3971881 T8026 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[201 (1440564861018832896)]} 0 1
   [junit4]   2> 3971887 T8055 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[201 (1440564861018832896)]} 0 1
   [junit4]   2> 3971889 T8023 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[201 (1440564861018832896)]} 0 12
   [junit4]   2> 3971923 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[202 (1440564861067067392)]} 0 1
   [junit4]   2> 3971941 T8072 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[202 (1440564861082796032)]} 0 1
   [junit4]   2> 3971942 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[202 (1440564861082796032)]} 0 5
   [junit4]   2> 3971944 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[202]} 0 10
   [junit4]   2> 3971983 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[203 (1440564861131030528)]} 0 1
   [junit4]   2> 3971994 T8024 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 203
   [junit4]   2> 3971999 T8025 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[203 (1440564861141516288)]} 0 0
   [junit4]   2> 3972004 T8056 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[203 (1440564861141516288)]} 0 1
   [junit4]   2> 3972006 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[203 (1440564861141516288)]} 0 13
   [junit4]   2> 3972008 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[203]} 0 20
   [junit4]   2> 3972047 T8006 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[191 (-1440564861197090816)]} 0 1
   [junit4]   2> 3972061 T8074 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[191 (-1440564861210722304)]} 0 0
   [junit4]   2> 3972063 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[191 (-1440564861210722304)]} 0 5
   [junit4]   2> 3972064 T8026 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[191]} 0 13
   [junit4]   2> 3972068 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[204 (1440564861219110912)]} 0 1
   [junit4]   2> 3972075 T8077 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[204 (1440564861223305216)]} 0 1
   [junit4]   2> 3972076 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[204 (1440564861223305216)]} 0 5
   [junit4]   2> 3972114 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[205 (1440564861267345408)]} 0 1
   [junit4]   2> 3972128 T8073 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[205 (1440564861277831168)]} 0 1
   [junit4]   2> 3972131 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[205 (1440564861277831168)]} 0 8
   [junit4]   2> 3972132 T8023 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[205]} 0 15
   [junit4]   2> 3972168 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[150 (-1440564861323968512)]} 0 1
   [junit4]   2> 3972180 T8028 C1838 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3972183 T8025 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[150 (-1440564861337600000)]} 0 0
   [junit4]   2> 3972188 T8024 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[150 (-1440564861337600000)]} 0 1
   [junit4]   2> 3972192 T8058 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[150 (-1440564861337600000)]} 0 1
   [junit4]   2> 3972193 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[150 (-1440564861337600000)]} 0 14
   [junit4]   2> 3972194 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[150]} 0 18
   [junit4]   2> 3972198 T8004 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[206 (1440564861355425792)]} 0 1
   [junit4]   2> 3972211 T8026 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 206
   [junit4]   2> 3972224 T8023 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[206 (1440564861369057280)]} 0 1
   [junit4]   2> 3972229 T8053 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[206 (1440564861369057280)]} 0 1
   [junit4]   2> 3972230 T8026 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[206 (1440564861369057280)]} 0 20
   [junit4]   2> 3972231 T8041 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[206]} 0 25
   [junit4]   2> 3972264 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[120 (-1440564861425680384)]} 0 0
   [junit4]   2> 3972274 T8076 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[120 (-1440564861433020416)]} 0 0
   [junit4]   2> 3972276 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[120 (-1440564861433020416)]} 0 5
   [junit4]   2> 3972277 T8025 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[120]} 0 9
   [junit4]   2> 3972281 T8006 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[207 (1440564861442457600)]} 0 1
   [junit4]   2> 3972292 T8028 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 207
   [junit4]   2> 3972308 T8023 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[207 (1440564861455040512)]} 0 1
   [junit4]   2> 3972313 T8059 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[207 (1440564861455040512)]} 0 1
   [junit4]   2> 3972314 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[207 (1440564861455040512)]} 0 22
   [junit4]   2> 3972316 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[207]} 0 30
   [junit4]   2> 3972351 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[194 (-1440564861516906496)]} 0 1
   [junit4]   2> 3972362 T8072 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[194 (-1440564861525295104)]} 0 0
   [junit4]   2> 3972364 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[194 (-1440564861525295104)]} 0 5
   [junit4]   2> 3972369 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[208 (1440564861534732288)]} 0 1
   [junit4]   2> 3972377 T8074 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[208 (1440564861539975168)]} 0 1
   [junit4]   2> 3972378 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[208 (1440564861539975168)]} 0 5
   [junit4]   2> 3972414 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[209 (1440564861581918208)]} 0 1
   [junit4]   2> 3972425 T8077 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[209 (1440564861589258240)]} 0 0
   [junit4]   2> 3972430 T8041 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[209 (1440564861589258240)]} 0 10
   [junit4]   2> 3972465 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[210 (1440564861635395584)]} 0 2
   [junit4]   2> 3972474 T8073 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[210 (1440564861641687040)]} 0 1
   [junit4]   2> 3972476 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[210 (1440564861641687040)]} 0 6
   [junit4]   2> 3972516 T8006 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[210 (-1440564861689921536)]} 0 1
   [junit4]   2> 3972524 T8075 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[210 (-1440564861694115840)]} 0 1
   [junit4]   2> 3972536 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[210 (-1440564861694115840)]} 0 16
   [junit4]   2> 3972546 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[211 (1440564861720330240)]} 0 1
   [junit4]   2> 3972560 T8026 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 211
   [junit4]   2> 3972564 T8025 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[211 (1440564861735010304)]} 0 1
   [junit4]   2> 3972568 T8055 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[211 (1440564861735010304)]} 0 1
   [junit4]   2> 3972569 T8026 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[211 (1440564861735010304)]} 0 10
   [junit4]   2> 3972605 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[212 (1440564861783244800)]} 0 1
   [junit4]   2> 3972623 T8024 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 212
   [junit4]   2> 3972630 T8023 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[212 (1440564861797924864)]} 0 0
   [junit4]   2> 3972634 T8056 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[212 (1440564861797924864)]} 0 0
   [junit4]   2> 3972636 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[212 (1440564861797924864)]} 0 17
   [junit4]   2> 3972637 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[212]} 0 24
   [junit4]   2> 3972678 T8004 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[167 (-1440564861859790848)]} 0 0
   [junit4]   2> 3972692 T8028 C1838 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3972699 T8026 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[167 (-1440564861873422336)]} 0 1
   [junit4]   2> 3972703 T8023 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[167 (-1440564861873422336)]} 0 1
   [junit4]   2> 3972710 T8058 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[167 (-1440564861873422336)]} 0 0
   [junit4]   2> 3972720 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[167 (-1440564861873422336)]} 0 29
   [junit4]   2> 3972729 T8041 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[167]} 0 41
   [junit4]   2> 3972734 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[213 (1440564861917462528)]} 0 1
   [junit4]   2> 3972746 T8024 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 213
   [junit4]   2> 3972751 T8025 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[213 (1440564861931094016)]} 0 1
   [junit4]   2> 3972755 T8053 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[213 (1440564861931094016)]} 0 0
   [junit4]   2> 3972757 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[213 (1440564861931094016)]} 0 11
   [junit4]   2> 3972801 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[214 (1440564861988765696)]} 0 1
   [junit4]   2> 3972806 T8014 oasc.OverseerCollectionProcessor.splitShard Asking sub shard leader to wait for: collection1_shard1_0_replica2 to be alive on: 127.0.0.1:51258_qb
   [junit4]   2> 3972807 T8014 oasc.OverseerCollectionProcessor.splitShard Creating replica shard collection1_shard1_1_replica2 as part of slice shard1_1 of collection collection1 on 127.0.0.1:51251_qb
   [junit4]   2> 3972812 T8026 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node7, state: active, checkLive: true, onlyIfLeader: true
   [junit4]   2> 3972813 T8039 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1_shard1_0_replica2
   [junit4]   2> 3972815 T8076 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[214 (1440564861992960000)]} 0 1
   [junit4]   2> 3972818 T8039 oasc.SolrXMLCoresLocator.doPersist Persisted core descriptions to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/solr.xml
   [junit4]   2> 3972818 T8039 oasc.CoreContainer.create Creating SolrCore 'collection1_shard1_0_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1_shard1_0_replica2
   [junit4]   2> 3972819 T8039 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 3972819 T8006 oasc.ZkController.publish publishing core=collection1_shard1_1_replica2 state=down
   [junit4]   2> 3972820 T8006 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3972820 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[214 (1440564861992960000)]} 0 15
   [junit4]   2> 3972823 T8039 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3972824 T8039 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 3972826 T8006 oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 3972828 T8039 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1_shard1_0_replica2/'
   [junit4]   2> 3972855 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[197 (-1440564862044340224)]} 0 1
   [junit4]   2> 3972868 T8023 C1838 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3972878 T8028 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[197 (-1440564862059020288)]} 0 1
   [junit4]   2> 3972888 T8025 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[197 (-1440564862059020288)]} 0 1
   [junit4]   2> 3972904 T8059 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[197 (-1440564862059020288)]} 0 0
   [junit4]   2> 3972914 T8023 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[197 (-1440564862059020288)]} 0 46
   [junit4]   2> 3972915 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[197]} 0 49
   [junit4]   2> 3972928 T8004 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[215 (1440564862121934848)]} 0 1
   [junit4]   2> 3972936 T8024 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 215
   [junit4]   2> 3972940 T8028 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[215 (1440564862129274880)]} 0 1
   [junit4]   2> 3972953 T8057 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[215 (1440564862129274880)]} 0 1
   [junit4]   2> 3972955 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[215 (1440564862129274880)]} 0 20
   [junit4]   2> 3972999 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[216 (1440564862195335168)]} 0 1
   [junit4]   2> 3973030 T8074 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[216 (1440564862221549568)]} 0 1
   [junit4]   2> 3973035 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[216 (1440564862221549568)]} 0 12
   [junit4]   2> 3973046 T8025 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[216]} 0 32
   [junit4]   2> 3973053 T8039 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 3973091 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[217 (1440564862291804160)]} 0 1
   [junit4]   2> 3973105 T8077 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[217 (1440564862300192768)]} 0 1
   [junit4]   2> 3973107 T8041 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[217 (1440564862300192768)]} 0 9
   [junit4]   2> 3973150 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[218 (1440564862354718720)]} 0 0
   [junit4]   2> 3973176 T8039 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3973178 T8023 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 218
   [junit4]   2> 3973180 T8039 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3973187 T8024 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[218 (1440564862383030272)]} 0 0
   [junit4]   2> 3973195 T8055 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[218 (1440564862383030272)]} 0 1
   [junit4]   2> 3973196 T8023 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[218 (1440564862383030272)]} 0 19
   [junit4]   2> 3973197 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[218]} 0 29
   [junit4]   2> 3973205 T8039 oass.IndexSchema.readSchema [collection1_shard1_0_replica2] Schema name=test
   [junit4]   2> 3973235 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[219 (1440564862443847680)]} 0 1
   [junit4]   2> 3973246 T8025 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 219
   [junit4]   2> 3973254 T8028 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[219 (1440564862455382016)]} 0 1
   [junit4]   2> 3973259 T8058 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[219 (1440564862455382016)]} 0 1
   [junit4]   2> 3973260 T8025 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[219 (1440564862455382016)]} 0 15
   [junit4]   2> 3973293 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[174 (-1440564862504665088)]} 0 0
   [junit4]   2> 3973307 T8073 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[174 (-1440564862516199424)]} 0 1
   [junit4]   2> 3973308 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[174 (-1440564862516199424)]} 0 4
   [junit4]   2> 3973311 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[174]} 0 10
   [junit4]   2> 3973321 T8003 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[220 (1440564862531928064)]} 0 4
   [junit4]   2> 3973330 T8023 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 220
   [junit4]   2> 3973338 T8028 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[220 (1440564862542413824)]} 0 2
   [junit4]   2> 3973344 T8053 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[220 (1440564862542413824)]} 0 1
   [junit4]   2> 3973351 T8023 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[220 (1440564862542413824)]} 0 23
   [junit4]   2> 3973354 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[220]} 0 28
   [junit4]   2> 3973381 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3973391 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1_shard1_1_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3973396 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[221 (1440564862612668416)]} 0 1
   [junit4]   2> 3973415 T8025 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 221
   [junit4]   2> 3973426 T8024 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[221 (1440564862631542784)]} 0 1
   [junit4]   2> 3973440 T8059 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[221 (1440564862631542784)]} 0 1
   [junit4]   2> 3973445 T8025 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[221 (1440564862631542784)]} 0 31
   [junit4]   2> 3973452 T8041 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[221]} 0 43
   [junit4]   2> 3973460 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3973460 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3973460 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3973460 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3973460 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3973465 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3973495 T8004 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[222 (1440564862715428864)]} 0 1
   [junit4]   2> 3973511 T8028 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 222
   [junit4]   2> 3973529 T8023 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[222 (1440564862732206080)]} 0 1
   [junit4]   2> 3973534 T8057 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[222 (1440564862732206080)]} 0 0
   [junit4]   2> 3973544 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[222 (1440564862732206080)]} 0 35
   [junit4]   2> 3973559 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[222]} 0 59
   [junit4]   2> 3973600 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[223 (1440564862825529344)]} 0 1
   [junit4]   2> 3973621 T8072 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[223 (1440564862839160832)]} 0 1
   [junit4]   2> 3973624 T8044 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[223 (1440564862839160832)]} 0 13
   [junit4]   2> 3973663 T8007 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[224 (1440564862891589632)]} 0 0
   [junit4]   2> 3973674 T8076 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[224 (1440564862901026816)]} 0 0
   [junit4]   2> 3973681 T8040 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[224 (1440564862901026816)]} 0 10
   [junit4]   2> 3973682 T8024 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[224]} 0 13
   [junit4]   2> 3973722 T8005 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[146 (-1440564862953455616)]} 0 1
   [junit4]   2> 3973733 T8023 C1838 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3973736 T8028 C1840 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[146 (-1440564862966038528)]} 0 0
   [junit4]   2> 3973746 T8025 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[146 (-1440564862966038528)]} 0 0
   [junit4]   2> 3973750 T8056 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[146 (-1440564862966038528)]} 0 0
   [junit4]   2> 3973751 T8023 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[146 (-1440564862966038528)]} 0 18
   [junit4]   2> 3973755 T8004 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[225 (1440564862988058624)]} 0 1
   [junit4]   2> 3973767 T8074 C1844 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[225 (1440564862994350080)]} 0 4
   [junit4]   2> 3973776 T8043 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[225 (1440564862994350080)]} 0 15
   [junit4]   2> 3973809 T8014 oasc.OverseerCollectionProcessor.splitShard Asking sub shard leader to wait for: collection1_shard1_1_replica2 to be alive on: 127.0.0.1:51251_qb
   [junit4]   2> 3973809 T8008 C1839 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[226 (1440564863044681728)]} 0 1
   [junit4]   2> 3973813 T8024 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node8, state: active, checkLive: true, onlyIfLeader: true
   [junit4]   2> 3973816 T8028 C1838 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 226
   [junit4]   2> 3973820 T8025 C1841 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[226 (1440564863053070336)]} 0 1
   [junit4]   2> 3973827 T8006 oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1_shard1_1_replica2
   [junit4]   2> 3973830 T8006 oasc.SolrXMLCoresLocator.doPersist Persisted core descriptions to /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/solr.xml
   [junit4]   2> 3973831 T8006 oasc.CoreContainer.create Creating SolrCore 'collection1_shard1_1_replica2' using instanceDir: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1_shard1_1_replica2
   [junit4]   2> 3973831 T8006 oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 3973830 T8058 C1842 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[226 (1440564863053070336)]} 0 0
   [junit4]   2> 3973838 T8028 C1838 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[226 (1440564863053070336)]} 0 22
   [junit4]   2> 3973840 T8042 C1843 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[226]} 0 27
   [junit4]   2> 3973841 T8006 oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 3973842 T8006 oasc.ZkController.readConfigName Load collection config from:/collections/collection1
   [junit4]   2> 3973844 T8006 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1_shard1_1_replica2/'
   [junit4]   2> ASYNC  NEW_CORE C1845 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1845_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3973889 T8003 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[169 (-1440564863128567808)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1846 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1846_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3973905 T8023 C1846 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> ASYNC  NEW_CORE C1847 name=collection1_shard1_0_replica1 org.apache.solr.core.SolrCore@5c1f6bce url=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1 node=127.0.0.1:51255_qb C1847_STATE=coll:collection1 core:collection1_shard1_0_replica1 props:{state=active, core=collection1_shard1_0_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3973915 T8025 C1847 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[169 (-1440564863146393600)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1848 name=collection1_shard1_1_replica1 org.apache.solr.core.SolrCore@54aa02ed url=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1 node=127.0.0.1:51255_qb C1848_STATE=coll:collection1 core:collection1_shard1_1_replica1 props:{state=active, core=collection1_shard1_1_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3973919 T8028 C1848 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[169 (-1440564863146393600)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1849 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1849_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3973925 T8053 C1849 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[169 (-1440564863146393600)]} 0 1
   [junit4]   2> 3973926 T8023 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[169 (-1440564863146393600)]} 0 21
   [junit4]   2> ASYNC  NEW_CORE C1850 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1850_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3973930 T8044 C1850 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[169]} 0 33
   [junit4]   2> 3973936 T8007 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[227 (1440564863177850880)]} 0 1
   [junit4]   2> 3973951 T8025 C1846 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 227
   [junit4]   2> 3973954 T8028 C1847 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[227 (1440564863193579520)]} 0 0
   [junit4]   2> 3973962 T8059 C1849 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[227 (1440564863193579520)]} 0 1
   [junit4]   2> 3973966 T8025 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[227 (1440564863193579520)]} 0 16
   [junit4]   2> 3973977 T8040 C1850 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[227]} 0 31
   [junit4]   2> 3973982 T8006 oasc.SolrConfig.<init> Using Lucene MatchVersion: LUCENE_50
   [junit4]   2> 3974029 T8005 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[228 (1440564863274319872)]} 0 3
   [junit4]   2> 3974041 T8023 C1846 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 228
   [junit4]   2> 3974045 T8028 C1848 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[228 (1440564863287951360)]} 0 1
   [junit4]   2> 3974056 T8057 C1849 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[228 (1440564863287951360)]} 0 1
   [junit4]   2> 3974064 T8023 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[228 (1440564863287951360)]} 0 24
   [junit4]   2> 3974113 T8004 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[229 (1440564863362400256)]} 0 3
   [junit4]   2> ASYNC  NEW_CORE C1851 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1851_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3974133 T8075 C1851 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[229 (1440564863381274624)]} 0 1
   [junit4]   2> 3974139 T8006 oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 3974140 T8041 C1850 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[229 (1440564863381274624)]} 0 11
   [junit4]   2> 3974141 T8028 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[229]} 0 15
   [junit4]   2> 3974153 T8006 oass.IndexSchema.readSchema Reading Solr Schema from schema.xml
   [junit4]   2> 3974181 T8003 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[148 (-1440564863434752000)]} 0 0
   [junit4]   2> 3974194 T8023 C1846 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3974201 T8025 C1847 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[148 (-1440564863447334912)]} 0 0
   [junit4]   2> 3974214 T8006 oass.IndexSchema.readSchema [collection1_shard1_1_replica2] Schema name=test
   [junit4]   2> 3974214 T8028 C1848 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[148 (-1440564863447334912)]} 0 0
   [junit4]   2> 3974220 T8055 C1849 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[148 (-1440564863447334912)]} 0 1
   [junit4]   2> 3974222 T8023 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[148 (-1440564863447334912)]} 0 31
   [junit4]   2> 3974243 T8007 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[230 (1440564863500812288)]} 0 0
   [junit4]   2> 3974262 T8073 C1851 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[230 (1440564863514443776)]} 0 1
   [junit4]   2> 3974263 T8043 C1850 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[230 (1440564863514443776)]} 0 8
   [junit4]   2> 3974302 T8005 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[231 (1440564863562678272)]} 0 0
   [junit4]   2> 3974322 T8025 C1846 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 231
   [junit4]   2> 3974329 T8028 C1848 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[231 (1440564863583649792)]} 0 1
   [junit4]   2> 3974342 T8058 C1849 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[231 (1440564863583649792)]} 0 1
   [junit4]   2> 3974350 T8025 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[231 (1440564863583649792)]} 0 29
   [junit4]   2> 3974353 T8044 C1850 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[231]} 0 37
   [junit4]   2> 3974393 T8008 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[190 (-1440564863657050112)]} 0 1
   [junit4]   2> 3974405 T8023 C1846 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3974409 T8028 C1847 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[190 (-1440564863669633024)]} 0 1
   [junit4]   2> 3974419 T8025 C1848 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[190 (-1440564863669633024)]} 0 1
   [junit4]   2> 3974425 T8053 C1849 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[190 (-1440564863669633024)]} 0 1
   [junit4]   2> 3974426 T8023 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[190 (-1440564863669633024)]} 0 22
   [junit4]   2> 3974435 T8004 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[232 (1440564863702138880)]} 0 0
   [junit4]   2> 3974447 T8028 C1846 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 232
   [junit4]   2> 3974465 T8025 C1848 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[232 (1440564863713673216)]} 0 4
   [junit4]   2> 3974469 T8059 C1849 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[232 (1440564863713673216)]} 0 1
   [junit4]   2> 3974471 T8028 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[232 (1440564863713673216)]} 0 25
   [junit4]   2> 3974517 T8003 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[233 (1440564863787073536)]} 0 2
   [junit4]   2> 3974548 T8076 C1851 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[233 (1440564863812239360)]} 0 1
   [junit4]   2> 3974550 T8040 C1850 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[233 (1440564863812239360)]} 0 10
   [junit4]   2> 3974553 T8025 C1846 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[233]} 0 23
   [junit4]   2> 3974595 T8005 C1845 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[234 (1440564863868862464)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1852 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1852_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3974639 T8077 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[234 (1440564863892979712)]} 0 0
   [junit4]   2> ASYNC  NEW_CORE C1853 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1853_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3974649 T8041 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[234 (1440564863892979712)]} 0 33
   [junit4]   2> ASYNC  NEW_CORE C1854 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1854_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3974654 T8028 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[234]} 0 48
   [junit4]   2> ASYNC  NEW_CORE C1855 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1855_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3974698 T8008 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[204 (-1440564863977914368)]} 0 1
   [junit4]   2> 3974717 T8074 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[204 (-1440564863986302976)]} 0 0
   [junit4]   2> 3974724 T8042 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[204 (-1440564863986302976)]} 0 18
   [junit4]   2> 3974726 T8023 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[204]} 0 23
   [junit4]   2> 3974743 T8004 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[235 (1440564864024051712)]} 0 1
   [junit4]   2> 3974759 T8039 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 3974768 T8039 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 3974771 T8039 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3974779 T8072 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[235 (1440564864056557568)]} 0 0
   [junit4]   2> 3974781 T8044 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[235 (1440564864056557568)]} 0 8
   [junit4]   2> 3974787 T8025 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[235]} 0 24
   [junit4]   2> 3974788 T8039 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 3974823 T8007 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[236 (1440564864108986368)]} 0 1
   [junit4]   2> 3974829 T8028 C1854 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 236
   [junit4]   2> ASYNC  NEW_CORE C1856 name=collection1_shard1_1_replica1 org.apache.solr.core.SolrCore@54aa02ed url=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1 node=127.0.0.1:51255_qb C1856_STATE=coll:collection1 core:collection1_shard1_1_replica1 props:{state=active, core=collection1_shard1_1_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3974834 T8023 C1856 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[236 (1440564864114229248)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1857 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1857_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3974839 T8057 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[236 (1440564864114229248)]} 0 1
   [junit4]   2> 3974841 T8028 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[236 (1440564864114229248)]} 0 13
   [junit4]   2> 3974879 T8003 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[228 (-1440564864167706624)]} 0 0
   [junit4]   2> 3974919 T8025 C1854 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> ASYNC  NEW_CORE C1858 name=collection1_shard1_0_replica1 org.apache.solr.core.SolrCore@5c1f6bce url=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1 node=127.0.0.1:51255_qb C1858_STATE=coll:collection1 core:collection1_shard1_0_replica1 props:{state=active, core=collection1_shard1_0_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3974923 T8028 C1858 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[228 (-1440564864209649664)]} 0 0
   [junit4]   2> 3974928 T8023 C1856 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[228 (-1440564864209649664)]} 0 0
   [junit4]   2> 3974933 T8055 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[228 (-1440564864209649664)]} 0 1
   [junit4]   2> 3974934 T8025 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[228 (-1440564864209649664)]} 0 16
   [junit4]   2> 3974935 T8040 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[228]} 0 30
   [junit4]   2> 3974939 T8005 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[237 (1440564864230621184)]} 0 0
   [junit4]   2> 3974952 T8073 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[237 (1440564864240058368)]} 0 0
   [junit4]   2> 3974954 T8041 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[237 (1440564864240058368)]} 0 6
   [junit4]   2> 3974993 T8004 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[180 (-1440564864286195712)]} 0 1
   [junit4]   2> 3975015 T8076 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[180 (-1440564864295632896)]} 0 1
   [junit4]   2> 3975016 T8043 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[180 (-1440564864295632896)]} 0 15
   [junit4]   2> 3975021 T8007 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[238 (1440564864315555840)]} 0 1
   [junit4]   2> 3975028 T8028 C1854 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 238
   [junit4]   2> 3975033 T8023 C1856 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[238 (1440564864322895872)]} 0 1
   [junit4]   2> 3975039 T8058 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[238 (1440564864322895872)]} 0 2
   [junit4]   2> 3975040 T8028 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[238 (1440564864322895872)]} 0 14
   [junit4]   2> 3975055 T8039 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3975056 T8039 oasc.SolrCore.<init> [collection1_shard1_0_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1_shard1_0_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/
   [junit4]   2> 3975057 T8039 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3975060 T8039 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data
   [junit4]   2> 3975071 T8039 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index/
   [junit4]   2> 3975072 T8039 oasc.SolrCore.initIndex WARN [collection1_shard1_0_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1_shard1_0_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 3975073 T8039 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index
   [junit4]   2> 3975080 T8003 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[239 (1440564864378470400)]} 0 0
   [junit4]   2> 3975081 T8039 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@356d4ac1 lockFactory=org.apache.lucene.store.NativeFSLockFactory@3033ac10),segFN=segments_1,generation=1}
   [junit4]   2> 3975084 T8039 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3975095 T8077 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[239 (1440564864390004736)]} 0 1
   [junit4]   2> 3975097 T8042 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[239 (1440564864390004736)]} 0 7
   [junit4]   2> 3975101 T8039 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3975101 T8039 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3975102 T8039 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3975103 T8039 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3975107 T8039 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3975107 T8039 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3975108 T8039 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3975109 T8039 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3975109 T8039 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3975112 T8025 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[239]} 0 26
   [junit4]   2> 3975135 T8039 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3975140 T8039 oass.SolrIndexSearcher.<init> Opening Searcher@626992b1 main
   [junit4]   2> 3975142 T8039 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3975143 T8039 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3975148 T8102 oasc.SolrCore.registerSearcher [collection1_shard1_0_replica2] Registered new searcher Searcher@626992b1 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3975151 T8008 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[240 (1440564864451870720)]} 0 1
   [junit4]   2> 3975156 T8039 oasc.CoreContainer.registerCore registering core: collection1_shard1_0_replica2
   [junit4]   2> 3975157 T8039 oasc.ZkController.register Register replica - core:collection1_shard1_0_replica2 address:http://127.0.0.1:51258/qb collection:collection1 shard:shard1_0
   [junit4]   2> 3975157 T8039 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3975171 T8023 C1854 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 240
   [junit4]   2> 3975179 T8028 C1858 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[240 (1440564864473890816)]} 0 1
   [junit4]   2> 3975192 T8053 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[240 (1440564864473890816)]} 0 1
   [junit4]   2> 3975202 T8023 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[240 (1440564864473890816)]} 0 32
   [junit4]   2> 3975213 T8040 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[240]} 0 49
   [junit4]   2> 3975248 T8005 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[241 (1440564864554631168)]} 0 0
   [junit4]   2> 3975266 T8025 C1854 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 241
   [junit4]   2> 3975278 T8028 C1856 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[241 (1440564864573505536)]} 0 1
   [junit4]   2> 3975297 T8039 oasc.ZkController.register We are http://127.0.0.1:51258/qb/collection1_shard1_0_replica2/ and leader is http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/
   [junit4]   2> 3975298 T8039 oasc.ZkController.checkRecovery Core needs to recover:collection1_shard1_0_replica2
   [junit4]   2> 3975299 T8039 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> 3975300 T8059 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[241 (1440564864573505536)]} 0 0
   [junit4]   2> 3975305 T8025 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[241 (1440564864573505536)]} 0 39
   [junit4]   2> 3975308 T8041 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[241]} 0 56
   [junit4]   2> ASYNC  NEW_CORE C1859 name=collection1_shard1_0_replica2 org.apache.solr.core.SolrCore@200e6d1f url=http://127.0.0.1:51258/qb/collection1_shard1_0_replica2 node=127.0.0.1:51258_qb C1859_STATE=coll:collection1 core:collection1_shard1_0_replica2 props:{state=down, core=collection1_shard1_0_replica2, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb}
   [junit4]   2> 3975310 T8103 C1859 P51258 oasc.RecoveryStrategy.run Starting recovery process.  core=collection1_shard1_0_replica2 recoveringAfterStartup=true
   [junit4]   2> 3975316 T8103 C1859 P51258 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 3975316 T8103 C1859 P51258 oasc.ZkController.publish publishing core=collection1_shard1_0_replica2 state=recovering
   [junit4]   2> 3975317 T8103 C1859 P51258 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3975329 T8103 C1859 P51258 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3975347 T8004 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[242 (1440564864657391616)]} 0 1
   [junit4]   2> 3975358 T8028 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node7, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 3975359 T8039 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3975375 T8025 C1854 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 242
   [junit4]   2> 3975380 T8039 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={shard=shard1_0&name=collection1_shard1_0_replica2&action=CREATE&collection=collection1&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=4576 
   [junit4]   2> 3975383 T8023 C1858 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[242 (1440564864686751744)]} 0 1
   [junit4]   2> 3975388 T8056 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[242 (1440564864686751744)]} 0 1
   [junit4]   2> 3975397 T8025 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[242 (1440564864686751744)]} 0 23
   [junit4]   2> 3975398 T8043 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[242]} 0 40
   [junit4]   2> 3975439 T8003 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[209 (-1440564864754909184)]} 0 0
   [junit4]   2> 3975480 T8075 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[209 (-1440564864790560768)]} 0 0
   [junit4]   2> 3975486 T8042 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[209 (-1440564864790560768)]} 0 13
   [junit4]   2> 3975491 T8023 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[209]} 0 28
   [junit4]   2> 3975509 T8008 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[243 (1440564864827260928)]} 0 2
   [junit4]   2> 3975520 T8025 C1854 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 243
   [junit4]   2> 3975525 T8023 C1856 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[243 (1440564864839843840)]} 0 1
   [junit4]   2> 3975534 T8055 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[243 (1440564864839843840)]} 0 1
   [junit4]   2> 3975540 T8025 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[243 (1440564864839843840)]} 0 21
   [junit4]   2> 3975586 T8005 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[244 (1440564864909049856)]} 0 0
   [junit4]   2> 3975591 T8023 C1854 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 244
   [junit4]   2> 3975603 T8025 C1856 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[244 (1440564864914292736)]} 0 1
   [junit4]   2> 3975624 T8058 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[244 (1440564864914292736)]} 0 1
   [junit4]   2> 3975636 T8023 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[244 (1440564864914292736)]} 0 46
   [junit4]   2> 3975672 T8007 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[159 (-1440564864997130240)]} 0 3
   [junit4]   2> 3975696 T8072 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[159 (-1440564865011810304)]} 0 0
   [junit4]   2> 3975706 T8044 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[159 (-1440564865011810304)]} 0 22
   [junit4]   2> 3975733 T8004 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[245 (1440564865057947648)]} 0 6
   [junit4]   2> 3975744 T8025 C1854 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 245
   [junit4]   2> 3975747 T8023 C1856 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[245 (1440564865073676288)]} 0 0
   [junit4]   2> 3975752 T8053 C1857 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[245 (1440564865073676288)]} 0 1
   [junit4]   2> 3975762 T8025 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[245 (1440564865073676288)]} 0 19
   [junit4]   2> 3975803 T8003 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[119 (-1440564865136590848)]} 0 1
   [junit4]   2> 3975836 T8073 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[119 (-1440564865158610944)]} 0 1
   [junit4]   2> 3975842 T8041 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[119 (-1440564865158610944)]} 0 18
   [junit4]   2> 3975855 T8025 C1854 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[119]} 0 36
   [junit4]   2> 3975870 T8005 C1855 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[246 (1440564865205796864)]} 0 1
   [junit4]   2> 3975897 T8076 C1852 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[246 (1440564865214185472)]} 0 1
   [junit4]   2> 3975900 T8039 C1853 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[246 (1440564865214185472)]} 0 24
   [junit4]   2> 3975979 T8006 oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> ASYNC  NEW_CORE C1860 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1860_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3975981 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[219 (-1440564865323237376)]} 0 0
   [junit4]   2> 3975999 T8006 oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> ASYNC  NEW_CORE C1861 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1861_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3976002 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3976008 T8006 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> ASYNC  NEW_CORE C1862 name=collection1_shard1_0_replica1 org.apache.solr.core.SolrCore@5c1f6bce url=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1 node=127.0.0.1:51255_qb C1862_STATE=coll:collection1 core:collection1_shard1_0_replica1 props:{state=active, core=collection1_shard1_0_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3976011 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[219 (-1440564865345257472)]} 0 0
   [junit4]   2> ASYNC  NEW_CORE C1863 name=collection1_shard1_1_replica1 org.apache.solr.core.SolrCore@54aa02ed url=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1 node=127.0.0.1:51255_qb C1863_STATE=coll:collection1 core:collection1_shard1_1_replica1 props:{state=active, core=collection1_shard1_1_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3976016 T8025 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[219 (-1440564865345257472)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1864 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1864_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3976021 T8059 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[219 (-1440564865345257472)]} 0 0
   [junit4]   2> 3976022 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[219 (-1440564865345257472)]} 0 21
   [junit4]   2> ASYNC  NEW_CORE C1865 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1865_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3976023 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[219]} 0 31
   [junit4]   2> 3976029 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[247 (1440564865372520448)]} 0 1
   [junit4]   2> 3976050 T8006 oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> ASYNC  NEW_CORE C1866 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1866_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3976059 T8074 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[247 (1440564865391394816)]} 0 2
   [junit4]   2> 3976065 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[247 (1440564865391394816)]} 0 19
   [junit4]   2> 3976100 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[248 (1440564865446969344)]} 0 1
   [junit4]   2> 3976124 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[248 (1440564865459552256)]} 0 1
   [junit4]   2> 3976127 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[248 (1440564865459552256)]} 0 16
   [junit4]   2> 3976131 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[248]} 0 25
   [junit4]   2> 3976172 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[205 (-1440564865523515392)]} 0 1
   [junit4]   2> 3976181 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[205 (-1440564865529806848)]} 0 0
   [junit4]   2> 3976183 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[205 (-1440564865529806848)]} 0 5
   [junit4]   2> 3976184 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[205]} 0 9
   [junit4]   2> 3976190 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[249 (1440564865542389760)]} 0 0
   [junit4]   2> 3976200 T8073 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[249 (1440564865549729792)]} 0 0
   [junit4]   2> 3976203 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[249 (1440564865549729792)]} 0 5
   [junit4]   2> 3976204 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[249]} 0 10
   [junit4]   2> 3976210 T8006 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 3976211 T8006 oasc.SolrCore.<init> [collection1_shard1_1_replica2] Opening new SolrCore at /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1_shard1_1_replica2/, dataDir=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/
   [junit4]   2> 3976212 T8006 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@178fd0ca
   [junit4]   2> 3976214 T8006 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data
   [junit4]   2> 3976215 T8006 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index/
   [junit4]   2> 3976216 T8006 oasc.SolrCore.initIndex WARN [collection1_shard1_1_replica2] Solr index directory '/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1_shard1_1_replica2/data/index' doesn't exist. Creating new index...
   [junit4]   2> 3976216 T8006 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index
   [junit4]   2> 3976222 T8006 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7db675f5 lockFactory=org.apache.lucene.store.NativeFSLockFactory@380ad5c4),segFN=segments_1,generation=1}
   [junit4]   2> 3976223 T8006 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3976231 T8006 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 3976232 T8006 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 3976233 T8006 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 3976235 T8006 oasc.RequestHandlers.initHandlersFromConfig created standard: solr.StandardRequestHandler
   [junit4]   2> 3976236 T8006 oasc.RequestHandlers.initHandlersFromConfig created /get: solr.RealTimeGetHandler
   [junit4]   2> 3976237 T8006 oasc.RequestHandlers.initHandlersFromConfig adding lazy requestHandler: solr.ReplicationHandler
   [junit4]   2> 3976237 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[195 (-1440564865591672832)]} 0 0
   [junit4]   2> 3976238 T8006 oasc.RequestHandlers.initHandlersFromConfig created /replication: solr.ReplicationHandler
   [junit4]   2> 3976241 T8006 oasc.RequestHandlers.initHandlersFromConfig created /update: solr.UpdateRequestHandler
   [junit4]   2> 3976243 T8006 oasc.RequestHandlers.initHandlersFromConfig created /admin/: org.apache.solr.handler.admin.AdminHandlers
   [junit4]   2> 3976247 T8077 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[195 (-1440564865595867136)]} 0 0
   [junit4]   2> 3976252 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[195 (-1440564865595867136)]} 0 11
   [junit4]   2> 3976272 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[250 (1440564865628372992)]} 0 0
   [junit4]   2> 3976273 T8006 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 3976276 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 250
   [junit4]   2> 3976279 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[250 (1440564865631518720)]} 0 1
   [junit4]   2> 3976282 T8056 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[250 (1440564865631518720)]} 0 0
   [junit4]   2> 3976283 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[250 (1440564865631518720)]} 0 8
   [junit4]   2> 3976285 T8006 oass.SolrIndexSearcher.<init> Opening Searcher@7e4c0573 main
   [junit4]   2> 3976287 T8006 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 3976287 T8006 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 3976296 T8104 oasc.SolrCore.registerSearcher [collection1_shard1_1_replica2] Registered new searcher Searcher@7e4c0573 main{StandardDirectoryReader(segments_1:1)}
   [junit4]   2> 3976299 T8006 oasc.CoreContainer.registerCore registering core: collection1_shard1_1_replica2
   [junit4]   2> 3976299 T8006 oasc.ZkController.register Register replica - core:collection1_shard1_1_replica2 address:http://127.0.0.1:51251/qb collection:collection1 shard:shard1_1
   [junit4]   2> 3976299 T8006 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=10000&maxConnectionsPerHost=20&connTimeout=30000&socketTimeout=30000&retry=false
   [junit4]   2> 3976317 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[251 (1440564865675558912)]} 0 0
   [junit4]   2> 3976330 T8076 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[251 (1440564865684996096)]} 0 1
   [junit4]   2> 3976332 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[251 (1440564865684996096)]} 0 6
   [junit4]   2> 3976334 T8006 oasc.ZkController.register We are http://127.0.0.1:51251/qb/collection1_shard1_1_replica2/ and leader is http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/
   [junit4]   2> 3976334 T8006 oasc.ZkController.checkRecovery Core needs to recover:collection1_shard1_1_replica2
   [junit4]   2> 3976335 T8006 oasu.DefaultSolrCoreState.doRecovery Running recovery - first canceling any ongoing recovery
   [junit4]   2> ASYNC  NEW_CORE C1867 name=collection1_shard1_1_replica2 org.apache.solr.core.SolrCore@1562478b url=http://127.0.0.1:51251/qb/collection1_shard1_1_replica2 node=127.0.0.1:51251_qb C1867_STATE=coll:collection1 core:collection1_shard1_1_replica2 props:{state=down, core=collection1_shard1_1_replica2, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb}
   [junit4]   2> 3976336 T8105 C1867 P51251 oasc.RecoveryStrategy.run Starting recovery process.  core=collection1_shard1_1_replica2 recoveringAfterStartup=true
   [junit4]   2> 3976337 T8105 C1867 P51251 oasc.RecoveryStrategy.doRecovery ###### startupVersions=[]
   [junit4]   2> 3976338 T8105 C1867 P51251 oasc.ZkController.publish publishing core=collection1_shard1_1_replica2 state=recovering
   [junit4]   2> 3976338 T8105 C1867 P51251 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3976340 T8006 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3976342 T8105 C1867 P51251 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3976363 T8006 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={shard=shard1_1&name=collection1_shard1_1_replica2&action=CREATE&collection=collection1&wt=javabin&qt=/admin/cores&version=2} status=0 QTime=3545 
   [junit4]   2> 3976366 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[232 (-1440564865726939136)]} 0 1
   [junit4]   2> 3976380 T8025 oasha.CoreAdminHandler.handleWaitForStateAction Going to wait for coreNodeName: core_node8, state: recovering, checkLive: true, onlyIfLeader: true
   [junit4]   2> 3976385 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3976390 T8106 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[232 (-1440564865746862080)]} 0 1
   [junit4]   2> 3976399 T8107 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[232 (-1440564865746862080)]} 0 0
   [junit4]   2> 3976404 T8058 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[232 (-1440564865746862080)]} 0 0
   [junit4]   2> 3976406 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[232 (-1440564865746862080)]} 0 22
   [junit4]   2> 3976419 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[252 (1440564865782513664)]} 0 1
   [junit4]   2> 3976431 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[252 (1440564865790902272)]} 0 1
   [junit4]   2> 3976433 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[252 (1440564865790902272)]} 0 6
   [junit4]   2> 3976434 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[252]} 0 10
   [junit4]   2> 3976484 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[172 (-1440564865849622528)]} 0 1
   [junit4]   2> 3976491 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[172 (-1440564865854865408)]} 0 0
   [junit4]   2> 3976493 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[172 (-1440564865854865408)]} 0 5
   [junit4]   2> 3976498 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[253 (1440564865864302592)]} 0 1
   [junit4]   2> 3976503 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 253
   [junit4]   2> 3976507 T8107 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[253 (1440564865870594048)]} 0 0
   [junit4]   2> 3976512 T8057 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[253 (1440564865870594048)]} 0 1
   [junit4]   2> 3976514 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[253 (1440564865870594048)]} 0 11
   [junit4]   2> 3976518 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3976531 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_0",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1_shard1_0_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 3976561 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[254 (1440564865930362880)]} 0 1
   [junit4]   2> 3976569 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"recovering",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1_shard1_1_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3976574 T8073 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[254 (1440564865941897216)]} 0 1
   [junit4]   2> 3976575 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[254 (1440564865941897216)]} 0 5
   [junit4]   2> 3976577 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[254]} 0 9
   [junit4]   2> 3976599 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3976599 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3976599 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3976600 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3976600 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3976609 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3976610 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[243 (-1440564865982791680)]} 0 1
   [junit4]   2> 3976623 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3976629 T8023 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[243 (-1440564865996423168)]} 0 2
   [junit4]   2>  C1867_STATE=coll:collection1 core:collection1_shard1_1_replica2 props:{state=recovering, core=collection1_shard1_1_replica2, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb}
   [junit4]   2> 3976685 T8004 C1867 P51251 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7db675f5 lockFactory=org.apache.lucene.store.NativeFSLockFactory@380ad5c4),segFN=segments_1,generation=1}
   [junit4]   2> 3976687 T8004 C1867 P51251 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3976690 T8004 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[243 (-1440564865996423168)]} 0 35
   [junit4]   2> 3976692 T8023 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[243 (-1440564865996423168)]} 0 55
   [junit4]   2> 3976703 T8053 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[243 (-1440564865996423168)]} 0 1
   [junit4]   2> 3976704 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[243 (-1440564865996423168)]} 0 81
   [junit4]   2> 3976715 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[243]} 0 95
   [junit4]   2> 3976726 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[255 (1440564866104426496)]} 0 0
   [junit4]   2> 3976743 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 255
   [junit4]   2> 3976764 T8008 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[255 (1440564866121203712)]} 0 5
   [junit4]   2> 3976766 T8107 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[255 (1440564866121203712)]} 0 18
   [junit4]   2> 3976771 T8055 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[255 (1440564866121203712)]} 0 0
   [junit4]   2> 3976772 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[255 (1440564866121203712)]} 0 30
   [junit4]   2> 3976773 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[255]} 0 39
   [junit4]   2> 3976811 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[183 (-1440564866193555456)]} 0 1
   [junit4]   2> 3976824 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> ASYNC  NEW_CORE C1868 name=collection1_shard1_0_replica2 org.apache.solr.core.SolrCore@200e6d1f url=http://127.0.0.1:51258/qb/collection1_shard1_0_replica2 node=127.0.0.1:51258_qb C1868_STATE=coll:collection1 core:collection1_shard1_0_replica2 props:{state=recovering, core=collection1_shard1_0_replica2, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb}
   [junit4]   2> 3976850 T8040 C1868 P51258 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@356d4ac1 lockFactory=org.apache.lucene.store.NativeFSLockFactory@3033ac10),segFN=segments_1,generation=1}
   [junit4]   2> 3976852 T8040 C1868 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 3976854 T8040 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[183 (-1440564866207186944)]} 0 7
   [junit4]   2> 3976855 T8023 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[183 (-1440564866207186944)]} 0 22
   [junit4]   2> 3976859 T8023 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[183 (-1440564866207186944)]} 0 1
   [junit4]   2> 3976864 T8056 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[183 (-1440564866207186944)]} 0 0
   [junit4]   2> 3976865 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[183 (-1440564866207186944)]} 0 41
   [junit4]   2> 3976866 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[183]} 0 50
   [junit4]   2> 3976869 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[256 (1440564866254372864)]} 0 1
   [junit4]   2> 3976885 T8077 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[256 (1440564866264858624)]} 0 1
   [junit4]   2> 3976886 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[256 (1440564866264858624)]} 0 8
   [junit4]   2> 3976887 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[256]} 0 14
   [junit4]   2> 3976920 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[248 (-1440564866306801664)]} 0 1
   [junit4]   2> 3976928 T8076 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[248 (-1440564866314141696)]} 0 0
   [junit4]   2> 3976930 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[248 (-1440564866314141696)]} 0 5
   [junit4]   2> 3976932 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[248]} 0 9
   [junit4]   2> 3976937 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[257 (1440564866324627456)]} 0 1
   [junit4]   2> 3976955 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 257
   [junit4]   2> 3976962 T8007 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[257 (1440564866344550400)]} 0 1
   [junit4]   2> 3976967 T8107 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[257 (1440564866344550400)]} 0 9
   [junit4]   2> 3976972 T8058 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[257 (1440564866344550400)]} 0 1
   [junit4]   2> 3976973 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[257 (1440564866344550400)]} 0 19
   [junit4]   2> 3976975 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[257]} 0 34
   [junit4]   2> 3977010 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[181 (-1440564866401173504)]} 0 3
   [junit4]   2> 3977024 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[181 (-1440564866413756416)]} 0 1
   [junit4]   2> 3977025 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[181 (-1440564866413756416)]} 0 5
   [junit4]   2> 3977030 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[181]} 0 12
   [junit4]   2> 3977041 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[258 (1440564866434727936)]} 0 1
   [junit4]   2> 3977046 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 258
   [junit4]   2> 3977053 T8044 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[258 (1440564866438922240)]} 0 1
   [junit4]   2> 3977055 T8107 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[258 (1440564866438922240)]} 0 5
   [junit4]   2> 3977058 T8057 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[258 (1440564866438922240)]} 0 0
   [junit4]   2> 3977060 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[258 (1440564866438922240)]} 0 15
   [junit4]   2> 3977094 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[237 (-1440564866490302464)]} 0 1
   [junit4]   2> 3977103 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[237 (-1440564866496593920)]} 0 1
   [junit4]   2> 3977105 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[237 (-1440564866496593920)]} 0 5
   [junit4]   2> 3977108 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[259 (1440564866504982528)]} 0 1
   [junit4]   2> 3977115 T8073 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[259 (1440564866509176832)]} 0 1
   [junit4]   2> 3977117 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[259 (1440564866509176832)]} 0 6
   [junit4]   2> 3977151 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[260 (1440564866549022720)]} 0 1
   [junit4]   2> 3977161 T8074 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[260 (1440564866556362752)]} 0 1
   [junit4]   2> 3977162 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[260 (1440564866556362752)]} 0 5
   [junit4]   2> 3977195 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[261 (1440564866596208640)]} 0 0
   [junit4]   2> 3977205 T8077 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[261 (1440564866602500096)]} 0 1
   [junit4]   2> 3977206 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[261 (1440564866602500096)]} 0 5
   [junit4]   2> 3977207 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[261]} 0 8
   [junit4]   2> 3977240 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[225 (-1440564866643394560)]} 0 0
   [junit4]   2> 3977247 T8076 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[225 (-1440564866647588864)]} 0 1
   [junit4]   2> 3977248 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[225 (-1440564866647588864)]} 0 4
   [junit4]   2> 3977252 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[262 (1440564866655977472)]} 0 0
   [junit4]   2> 3977261 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[262 (1440564866661220352)]} 0 1
   [junit4]   2> 3977262 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[262 (1440564866661220352)]} 0 5
   [junit4]   2> 3977263 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[262]} 0 8
   [junit4]   2> 3977296 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[131 (-1440564866702114816)]} 0 0
   [junit4]   2> 3977305 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[131 (-1440564866708406272)]} 0 1
   [junit4]   2> 3977306 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[131 (-1440564866708406272)]} 0 4
   [junit4]   2> 3977307 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[131]} 0 8
   [junit4]   2> 3977311 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[263 (1440564866717843456)]} 0 0
   [junit4]   2> 3977319 T8074 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[263 (1440564866722037760)]} 0 1
   [junit4]   2> 3977320 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[263 (1440564866722037760)]} 0 5
   [junit4]   2> 3977355 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[247 (-1440564866762932224)]} 0 1
   [junit4]   2> 3977364 T8028 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node7, state: recovering, checkLive: true, onlyIfLeader: true for: 2 seconds.
   [junit4]   2> 3977365 T8028 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node7&state=recovering&nodeName=127.0.0.1:51258_qb&action=PREPRECOVERY&checkLive=true&core=collection1_shard1_0_replica1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=2007 
   [junit4]   2> 3977369 T8077 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[247 (-1440564866770272256)]} 0 0
   [junit4]   2> 3977372 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[247 (-1440564866770272256)]} 0 11
   [junit4]   2> 3977378 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[247]} 0 20
   [junit4]   2> 3977383 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[264 (1440564866792292352)]} 0 1
   [junit4]   2> 3977383 T8025 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node8, state: recovering, checkLive: true, onlyIfLeader: true for: 1 seconds.
   [junit4]   2> 3977385 T8025 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node8&state=recovering&nodeName=127.0.0.1:51251_qb&action=PREPRECOVERY&checkLive=true&core=collection1_shard1_1_replica1&wt=javabin&onlyIfLeader=true&version=2} status=0 QTime=1005 
   [junit4]   2> 3977394 T8076 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[264 (1440564866801729536)]} 0 1
   [junit4]   2> 3977395 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[264 (1440564866801729536)]} 0 4
   [junit4]   2> 3977397 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[264]} 0 10
   [junit4]   2> 3977432 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[264 (-1440564866843672576)]} 0 1
   [junit4]   2> 3977441 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[264 (-1440564866851012608)]} 0 0
   [junit4]   2> 3977443 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[264 (-1440564866851012608)]} 0 5
   [junit4]   2> 3977450 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[265 (1440564866863595520)]} 0 1
   [junit4]   2> 3977459 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 265
   [junit4]   2> 3977465 T8041 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[265 (1440564866871984128)]} 0 1
   [junit4]   2> 3977466 T8107 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[265 (1440564866871984128)]} 0 5
   [junit4]   2> 3977469 T8053 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[265 (1440564866871984128)]} 0 0
   [junit4]   2> 3977471 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[265 (1440564866871984128)]} 0 13
   [junit4]   2> 3977472 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[265]} 0 17
   [junit4]   2> 3977505 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[185 (-1440564866921267200)]} 0 0
   [junit4]   2> 3977511 T8028 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3977517 T8043 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[185 (-1440564866927558656)]} 0 0
   [junit4]   2> 3977518 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[185 (-1440564866927558656)]} 0 4
   [junit4]   2> 3977522 T8107 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[185 (-1440564866927558656)]} 0 1
   [junit4]   2> 3977525 T8055 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[185 (-1440564866927558656)]} 0 0
   [junit4]   2> 3977526 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[185 (-1440564866927558656)]} 0 16
   [junit4]   2> 3977530 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[266 (1440564866947481600)]} 0 0
   [junit4]   2> 3977537 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[266 (1440564866951675904)]} 0 0
   [junit4]   2> 3977539 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[266 (1440564866951675904)]} 0 5
   [junit4]   2> 3977575 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[267 (1440564866993618944)]} 0 1
   [junit4]   2> 3977581 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 267
   [junit4]   2> 3977587 T8044 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[267 (1440564866999910400)]} 0 0
   [junit4]   2> 3977588 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[267 (1440564866999910400)]} 0 4
   [junit4]   2> 3977595 T8056 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[267 (1440564866999910400)]} 0 1
   [junit4]   2> 3977596 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[267 (1440564866999910400)]} 0 16
   [junit4]   2> 3977597 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[267]} 0 19
   [junit4]   2> 3977631 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[268 (1440564867052339200)]} 0 1
   [junit4]   2> 3977640 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 268
   [junit4]   2> 3977651 T8041 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[268 (1440564867062824960)]} 0 1
   [junit4]   2> 3977653 T8028 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[268 (1440564867062824960)]} 0 7
   [junit4]   2> 3977658 T8058 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[268 (1440564867062824960)]} 0 1
   [junit4]   2> 3977659 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[268 (1440564867062824960)]} 0 20
   [junit4]   2> 3977696 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[269 (1440564867119448064)]} 0 3
   [junit4]   2> 3977706 T8073 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[269 (1440564867128885248)]} 0 0
   [junit4]   2> 3977708 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[269 (1440564867128885248)]} 0 5
   [junit4]   2> 3977741 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[230 (-1440564867168731136)]} 0 0
   [junit4]   2> 3977750 T8077 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[230 (-1440564867175022592)]} 0 0
   [junit4]   2> 3977751 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[230 (-1440564867175022592)]} 0 4
   [junit4]   2> 3977752 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[230]} 0 8
   [junit4]   2> 3977760 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[270 (1440564867187605504)]} 0 3
   [junit4]   2> 3977773 T8076 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[270 (1440564867198091264)]} 0 1
   [junit4]   2> 3977774 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[270 (1440564867198091264)]} 0 5
   [junit4]   2> 3977775 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[270]} 0 9
   [junit4]   2> 3977809 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[271 (1440564867238985728)]} 0 1
   [junit4]   2> 3977815 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[271 (1440564867243180032)]} 0 0
   [junit4]   2> 3977819 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[271 (1440564867243180032)]} 0 7
   [junit4]   2> 3977856 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[255 (-1440564867289317376)]} 0 1
   [junit4]   2> 3977860 T8028 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3977863 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[255 (-1440564867292463104)]} 0 0
   [junit4]   2> 3977869 T8005 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[255 (-1440564867292463104)]} 0 1
   [junit4]   2> 3977870 T8023 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[255 (-1440564867292463104)]} 0 4
   [junit4]   2> 3977874 T8057 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[255 (-1440564867292463104)]} 0 1
   [junit4]   2> 3977875 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[255 (-1440564867292463104)]} 0 16
   [junit4]   2> 3977878 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[272 (1440564867312386048)]} 0 1
   [junit4]   2> 3977884 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[272 (1440564867315531776)]} 0 0
   [junit4]   2> 3977885 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[272 (1440564867315531776)]} 0 4
   [junit4]   2> 3977919 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[273 (1440564867354329088)]} 0 1
   [junit4]   2> 3977922 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 273
   [junit4]   2> 3977931 T8039 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[273 (1440564867358523392)]} 0 1
   [junit4]   2> 3977934 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[273 (1440564867358523392)]} 0 8
   [junit4]   2> 3977941 T8053 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[273 (1440564867358523392)]} 0 1
   [junit4]   2> 3977943 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[273 (1440564867358523392)]} 0 21
   [junit4]   2> 3977978 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[274 (1440564867416195072)]} 0 1
   [junit4]   2> 3977988 T8074 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[274 (1440564867424583680)]} 0 1
   [junit4]   2> 3977990 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[274 (1440564867424583680)]} 0 5
   [junit4]   2> 3977991 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[274]} 0 9
   [junit4]   2> 3978029 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[137 (-1440564867470721024)]} 0 1
   [junit4]   2> 3978042 T8073 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[137 (-1440564867480158208)]} 0 1
   [junit4]   2> 3978045 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[137 (-1440564867480158208)]} 0 7
   [junit4]   2> 3978048 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[137]} 0 13
   [junit4]   2> 3978055 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[275 (1440564867496935424)]} 0 2
   [junit4]   2> 3978060 T8025 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 275
   [junit4]   2> 3978066 T8042 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[275 (1440564867502178304)]} 0 1
   [junit4]   2> 3978067 T8107 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[275 (1440564867502178304)]} 0 5
   [junit4]   2> 3978071 T8055 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[275 (1440564867502178304)]} 0 0
   [junit4]   2> 3978072 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[275 (1440564867502178304)]} 0 13
   [junit4]   2> 3978106 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[276 (1440564867550412800)]} 0 1
   [junit4]   2> 3978114 T8077 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[276 (1440564867556704256)]} 0 1
   [junit4]   2> 3978121 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[276 (1440564867556704256)]} 0 11
   [junit4]   2> 3978156 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[277 (1440564867602841600)]} 0 1
   [junit4]   2> 3978161 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 277
   [junit4]   2> 3978168 T8004 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[277 (1440564867608084480)]} 0 0
   [junit4]   2> 3978170 T8028 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[277 (1440564867608084480)]} 0 7
   [junit4]   2> 3978173 T8056 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[277 (1440564867608084480)]} 0 0
   [junit4]   2> 3978174 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[277 (1440564867608084480)]} 0 14
   [junit4]   2> 3978210 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[199 (-1440564867660513280)]} 0 0
   [junit4]   2> 3978218 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[199 (-1440564867666804736)]} 0 0
   [junit4]   2> 3978220 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[199 (-1440564867666804736)]} 0 4
   [junit4]   2> 3978221 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[199]} 0 8
   [junit4]   2> 3978225 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[278 (1440564867675193344)]} 0 1
   [junit4]   2> 3978239 T8025 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 278
   [junit4]   2> 3978249 T8008 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[278 (1440564867689873408)]} 0 2
   [junit4]   2> 3978258 T8028 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[278 (1440564867689873408)]} 0 16
   [junit4]   2> 3978275 T8058 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[278 (1440564867689873408)]} 0 1
   [junit4]   2> 3978277 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[278 (1440564867689873408)]} 0 39
   [junit4]   2> 3978278 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[278]} 0 43
   [junit4]   2> 3978311 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[262 (-1440564867765370880)]} 0 1
   [junit4]   2> 3978325 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[262 (-1440564867777953792)]} 0 0
   [junit4]   2> 3978326 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[262 (-1440564867777953792)]} 0 5
   [junit4]   2> 3978328 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[262]} 0 14
   [junit4]   2> 3978332 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[279 (1440564867788439552)]} 0 1
   [junit4]   2> 3978348 T8074 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[279 (1440564867798925312)]} 0 1
   [junit4]   2> 3978359 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[279 (1440564867798925312)]} 0 17
   [junit4]   2> 3978408 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[280 (1440564867867082752)]} 0 1
   [junit4]   2> 3978415 T8028 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 280
   [junit4]   2> 3978424 T8003 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[280 (1440564867875471360)]} 0 1
   [junit4]   2> 3978425 T8023 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[280 (1440564867875471360)]} 0 6
   [junit4]   2> 3978430 T8057 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[280 (1440564867875471360)]} 0 1
   [junit4]   2> 3978432 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[280 (1440564867875471360)]} 0 18
   [junit4]   2> 3978433 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[280]} 0 21
   [junit4]   2> 3978467 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[188 (-1440564867928948736)]} 0 1
   [junit4]   2> 3978471 T8025 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3978479 T8107 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[188 (-1440564867934191616)]} 0 3
   [junit4]   2> 3978491 T8005 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[188 (-1440564867934191616)]} 0 1
   [junit4]   2> 3978493 T8023 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[188 (-1440564867934191616)]} 0 6
   [junit4]   2> 3978498 T8053 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[188 (-1440564867934191616)]} 0 0
   [junit4]   2> 3978499 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[188 (-1440564867934191616)]} 0 28
   [junit4]   2> 3978503 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[281 (1440564867967746048)]} 0 1
   [junit4]   2> 3978512 T8073 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[281 (1440564867974037504)]} 0 1
   [junit4]   2> 3978514 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[281 (1440564867974037504)]} 0 6
   [junit4]   2> 3978516 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[281]} 0 10
   [junit4]   2> 3978549 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[282 (1440564868015980544)]} 0 0
   [junit4]   2> 3978561 T8077 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[282 (1440564868021223424)]} 0 2
   [junit4]   2> 3978566 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[282 (1440564868021223424)]} 0 13
   [junit4]   2> 3978604 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[283 (1440564868072603648)]} 0 1
   [junit4]   2> 3978612 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 283
   [junit4]   2> 3978621 T8003 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[283 (1440564868080992256)]} 0 1
   [junit4]   2> 3978622 T8023 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[283 (1440564868080992256)]} 0 5
   [junit4]   2> 3978637 T8055 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[283 (1440564868080992256)]} 0 5
   [junit4]   2> 3978643 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[283 (1440564868080992256)]} 0 32
   [junit4]   2> 3978645 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[283]} 0 37
   [junit4]   2> 3978678 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[253 (-1440564868151246848)]} 0 0
   [junit4]   2> 3978686 T8025 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3978697 T8042 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[253 (-1440564868158586880)]} 0 1
   [junit4]   2> 3978698 T8023 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[253 (-1440564868158586880)]} 0 6
   [junit4]   2> 3978703 T8107 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[253 (-1440564868158586880)]} 0 0
   [junit4]   2> 3978708 T8056 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[253 (-1440564868158586880)]} 0 1
   [junit4]   2> 3978709 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[253 (-1440564868158586880)]} 0 25
   [junit4]   2> 3978713 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[284 (1440564868187947008)]} 0 1
   [junit4]   2> 3978718 T8028 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 284
   [junit4]   2> 3978725 T8004 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[284 (1440564868193189888)]} 0 1
   [junit4]   2> 3978727 T8023 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[284 (1440564868193189888)]} 0 6
   [junit4]   2> 3978731 T8059 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[284 (1440564868193189888)]} 0 0
   [junit4]   2> 3978732 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[284 (1440564868193189888)]} 0 15
   [junit4]   2> 3978768 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[285 (1440564868244570112)]} 0 1
   [junit4]   2> 3978775 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 285
   [junit4]   2> 3978783 T8003 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[285 (1440564868251910144)]} 0 1
   [junit4]   2> 3978785 T8025 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[285 (1440564868251910144)]} 0 7
   [junit4]   2> 3978788 T8057 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[285 (1440564868251910144)]} 0 0
   [junit4]   2> 3978789 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[285 (1440564868251910144)]} 0 15
   [junit4]   2> 3978790 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[285]} 0 19
   [junit4]   2> 3978826 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[286 (1440564868306436096)]} 0 1
   [junit4]   2> 3978837 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[286 (1440564868313776128)]} 0 1
   [junit4]   2> 3978838 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[286 (1440564868313776128)]} 0 6
   [junit4]   2> 3978840 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[286]} 0 10
   [junit4]   2> 3978874 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[287 (1440564868355719168)]} 0 1
   [junit4]   2> 3978881 T8028 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 287
   [junit4]   2> 3978894 T8043 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[287 (1440564868363059200)]} 0 1
   [junit4]   2> 3978895 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[287 (1440564868363059200)]} 0 9
   [junit4]   2> 3978900 T8053 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[287 (1440564868363059200)]} 0 1
   [junit4]   2> 3978901 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[287 (1440564868363059200)]} 0 21
   [junit4]   2> 3978902 T8041 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[287]} 0 25
   [junit4]   2> 3978939 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[211 (-1440564868424925184)]} 0 0
   [junit4]   2> 3978943 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3978946 T8023 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[211 (-1440564868429119488)]} 0 0
   [junit4]   2> 3978952 T8004 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[211 (-1440564868429119488)]} 0 0
   [junit4]   2> 3978956 T8025 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[211 (-1440564868429119488)]} 0 6
   [junit4]   2> 3978963 T8055 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[211 (-1440564868429119488)]} 0 0
   [junit4]   2> 3978969 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[211 (-1440564868429119488)]} 0 27
   [junit4]   2> 3978974 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[288 (1440564868460576768)]} 0 1
   [junit4]   2> 3978978 T8028 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 288
   [junit4]   2> 3978985 T8040 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[288 (1440564868465819648)]} 0 1
   [junit4]   2> 3978986 T8023 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[288 (1440564868465819648)]} 0 5
   [junit4]   2> 3978990 T8058 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[288 (1440564868465819648)]} 0 1
   [junit4]   2> 3978992 T8028 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[288 (1440564868465819648)]} 0 15
   [junit4]   2> 3979033 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[289 (1440564868520345600)]} 0 5
   [junit4]   2> 3979044 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[289 (1440564868530831360)]} 0 1
   [junit4]   2> 3979046 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[289 (1440564868530831360)]} 0 6
   [junit4]   2> 3979048 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[289]} 0 11
   [junit4]   2> 3979081 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[269 (-1440564868572774400)]} 0 1
   [junit4]   2> 3979087 T8073 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[269 (-1440564868578017280)]} 0 0
   [junit4]   2> 3979089 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[269 (-1440564868578017280)]} 0 5
   [junit4]   2> 3979097 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[290 (1440564868588503040)]} 0 3
   [junit4]   2> 3979107 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 290
   [junit4]   2> 3979117 T8041 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[290 (1440564868600037376)]} 0 1
   [junit4]   2> 3979118 T8028 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[290 (1440564868600037376)]} 0 8
   [junit4]   2> 3979123 T8056 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[290 (1440564868600037376)]} 0 1
   [junit4]   2> 3979124 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[290 (1440564868600037376)]} 0 18
   [junit4]   2> 3979126 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[290]} 0 23
   [junit4]   2> 3979160 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[214 (-1440564868655611904)]} 0 1
   [junit4]   2> 3979170 T8076 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[214 (-1440564868659806208)]} 0 2
   [junit4]   2> 3979174 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[214 (-1440564868659806208)]} 0 11
   [junit4]   2> 3979181 T8007 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[291 (1440564868678680576)]} 0 1
   [junit4]   2> 3979193 T8077 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[291 (1440564868686020608)]} 0 0
   [junit4]   2> 3979195 T8042 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[291 (1440564868686020608)]} 0 7
   [junit4]   2> 3979229 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[292 (1440564868727963648)]} 0 1
   [junit4]   2> 3979238 T8075 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[292 (1440564868735303680)]} 0 1
   [junit4]   2> 3979239 T8044 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[292 (1440564868735303680)]} 0 4
   [junit4]   2> 3979241 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[292]} 0 9
   [junit4]   2> 3979277 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[293 (1440564868778295296)]} 0 1
   [junit4]   2> 3979281 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 293
   [junit4]   2> 3979290 T8039 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[293 (1440564868782489600)]} 0 0
   [junit4]   2> 3979292 T8028 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[293 (1440564868782489600)]} 0 5
   [junit4]   2> 3979296 T8057 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[293 (1440564868782489600)]} 0 1
   [junit4]   2> 3979298 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[293 (1440564868782489600)]} 0 18
   [junit4]   2> 3979331 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[294 (1440564868834918400)]} 0 1
   [junit4]   2> 3979335 T8107 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 294
   [junit4]   2> 3979341 T8041 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[294 (1440564868839112704)]} 0 1
   [junit4]   2> 3979343 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[294 (1440564868839112704)]} 0 6
   [junit4]   2> 3979347 T8053 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[294 (1440564868839112704)]} 0 1
   [junit4]   2> 3979348 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[294 (1440564868839112704)]} 0 14
   [junit4]   2> 3979370 T8103 C1868 P51258 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ core=collection1_shard1_0_replica2 - recoveringAfterStartup=true
   [junit4]   2> 3979370 T8103 C1868 P51258 oasu.PeerSync.sync PeerSync: core=collection1_shard1_0_replica2 url=http://127.0.0.1:51258/qb START replicas=[http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/] nUpdates=100
   [junit4]   2> 3979371 T8103 C1868 P51258 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 3979371 T8103 C1868 P51258 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=collection1_shard1_0_replica2
   [junit4]   2> 3979371 T8103 C1868 P51258 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=collection1_shard1_0_replica2
   [junit4]   2> 3979372 T8103 C1868 P51258 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=collection1_shard1_0_replica2
   [junit4]   2> 3979372 T8103 C1868 P51258 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=tlog{file=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1_shard1_0_replica2/data/tlog/tlog.0000000000000000000 refcount=1}}
   [junit4]   2> 3979372 T8103 C1868 P51258 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/. core=collection1_shard1_0_replica2
   [junit4]   2> 3979372 T8103 C1868 P51258 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3979377 T8028 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=0 
   [junit4]   2> 3979387 T8105 C1867 P51251 oasc.RecoveryStrategy.doRecovery Attempting to PeerSync from http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ core=collection1_shard1_1_replica2 - recoveringAfterStartup=true
   [junit4]   2> 3979388 T8105 C1867 P51251 oasu.PeerSync.sync PeerSync: core=collection1_shard1_1_replica2 url=http://127.0.0.1:51251/qb START replicas=[http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] nUpdates=100
   [junit4]   2> 3979392 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[295 (1440564868898881536)]} 0 1
   [junit4]   2> 3979393 T8105 C1867 P51251 oasu.PeerSync.sync WARN no frame of reference to tell of we've missed updates
   [junit4]   2> 3979396 T8105 C1867 P51251 oasc.RecoveryStrategy.doRecovery PeerSync Recovery was not successful - trying replication. core=collection1_shard1_1_replica2
   [junit4]   2> 3979396 T8105 C1867 P51251 oasc.RecoveryStrategy.doRecovery Starting Replication Recovery. core=collection1_shard1_1_replica2
   [junit4]   2> 3979396 T8105 C1867 P51251 oasc.RecoveryStrategy.doRecovery Begin buffering updates. core=collection1_shard1_1_replica2
   [junit4]   2> 3979397 T8025 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/get params={getVersions=100&distrib=false&wt=javabin&qt=/get&version=2} status=0 QTime=1 
   [junit4]   2> 3979395 T8023 C1862 P51255 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3979397 T8105 C1867 P51251 oasu.UpdateLog.bufferUpdates Starting to buffer updates. FSUpdateLog{state=ACTIVE, tlog=tlog{file=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1_shard1_1_replica2/data/tlog/tlog.0000000000000000000 refcount=1}}
   [junit4]   2> 3979409 T8023 C1862 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_2,generation=2}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_3,generation=3}
   [junit4]   2> 3979412 T8105 C1867 P51251 oasc.RecoveryStrategy.replicate Attempting to replicate from http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/. core=collection1_shard1_1_replica2
   [junit4]   2> 3979414 T8023 C1862 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3979414 T8105 C1867 P51251 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3979419 T8074 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[295 (1440564868924047360)]} 0 1
   [junit4]   2> 3979421 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[295 (1440564868924047360)]} 0 6
   [junit4]   2> 3979423 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[295]} 0 11
   [junit4]   2> 3979426 T8023 C1862 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@31b8ec1e realtime
   [junit4]   2> 3979427 T8023 C1862 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3979429 T8023 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 34
   [junit4]   2> 3979431 T8103 C1868 P51258 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 3979431 T8103 C1868 P51258 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 3979438 T8028 C1862 P51255 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 3979439 T8028 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=2 
   [junit4]   2> 3979443 T8103 C1868 P51258 oash.SnapPuller.fetchLatestIndex Master's generation: 3
   [junit4]   2> 3979443 T8103 C1868 P51258 oash.SnapPuller.fetchLatestIndex Slave's generation: 1
   [junit4]   2> 3979443 T8103 C1868 P51258 oash.SnapPuller.fetchLatestIndex Starting replication process
   [junit4]   2> 3979447 T8023 C1863 P51255 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=false,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3979450 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={command=filelist&qt=/replication&wt=javabin&generation=3&version=2} status=0 QTime=4 
   [junit4]   2> 3979452 T8103 C1868 P51258 oash.SnapPuller.fetchLatestIndex Number of files in latest index in master: 25
   [junit4]   2> 3979454 T8103 C1868 P51258 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index.20130714202156647
   [junit4]   2> 3979455 T8023 C1863 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_2,generation=2}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_3,generation=3}
   [junit4]   2> 3979456 T8023 C1863 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3979457 T8103 C1868 P51258 oash.SnapPuller.fetchLatestIndex Starting download to MockDirWrapper(org.apache.lucene.store.RAMDirectory@2f254955 lockFactory=org.apache.lucene.store.NativeFSLockFactory@23e45ac7) fullCopy=true
   [junit4]   2> 3979461 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[296 (1440564868972281856)]} 0 1
   [junit4]   2> 3979467 T8028 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_2.si&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979469 T8025 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 296
   [junit4]   2> 3979489 T8023 C1863 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@55cd0114 realtime
   [junit4]   2> 3979490 T8023 C1863 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3979491 T8023 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={waitSearcher=true&openSearcher=false&commit=true&wt=javabin&commit_end_point=true&version=2&softCommit=false} {commit=} 0 44
   [junit4]   2> 3979494 T8023 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_1_Lucene41_0.doc&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979495 T8105 C1867 P51251 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 3979495 T8105 C1867 P51251 oash.SnapPuller.<init>  No value set for 'pollInterval'. Timer Task not started.
   [junit4]   2> 3979506 T8007 C1867 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[296 (1440564868979621888)]} 0 1
   [junit4]   2> 3979508 T8107 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[296 (1440564868979621888)]} 0 36
   [junit4]   2> 3979510 T8028 C1863 P51255 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 3979510 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_2_Lucene41_0.tim&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979511 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={command=indexversion&qt=/replication&wt=javabin&version=2} status=0 QTime=4 
   [junit4]   2> 3979515 T8105 C1867 P51251 oash.SnapPuller.fetchLatestIndex Master's generation: 3
   [junit4]   2> 3979516 T8105 C1867 P51251 oash.SnapPuller.fetchLatestIndex Slave's generation: 1
   [junit4]   2> 3979516 T8105 C1867 P51251 oash.SnapPuller.fetchLatestIndex Starting replication process
   [junit4]   2> 3979518 T8055 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[296 (1440564868979621888)]} 0 1
   [junit4]   2> 3979519 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[296 (1440564868979621888)]} 0 51
   [junit4]   2> 3979521 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[296]} 0 55
   [junit4]   2> 3979527 T8023 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={command=filelist&qt=/replication&wt=javabin&generation=3&version=2} status=0 QTime=7 
   [junit4]   2> 3979529 T8105 C1867 P51251 oash.SnapPuller.fetchLatestIndex Number of files in latest index in master: 25
   [junit4]   2> 3979530 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_2_Lucene41_0.tip&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979532 T8105 C1867 P51251 oasc.CachingDirectoryFactory.get return new directory for /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index.20130714202156725
   [junit4]   2> 3979535 T8105 C1867 P51251 oash.SnapPuller.fetchLatestIndex Starting download to MockDirWrapper(org.apache.lucene.store.RAMDirectory@22ec99d4 lockFactory=org.apache.lucene.store.NativeFSLockFactory@22d86580) fullCopy=true
   [junit4]   2> 3979542 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_2.si&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979542 T8025 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_1.fnm&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979549 T8023 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_1_Lucene41_0.doc&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979549 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_0.si&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979554 T8008 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[297 (1440564869069799424)]} 0 1
   [junit4]   2> 3979559 T8025 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 297
   [junit4]   2> 3979560 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_2_Lucene41_0.tim&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979561 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_0_Lucene41_0.tim&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979566 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_2_Lucene41_0.tip&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979569 T8044 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[297 (1440564869073993728)]} 0 0
   [junit4]   2> 3979567 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_2_1.del&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979570 T8023 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[297 (1440564869073993728)]} 0 7
   [junit4]   2> 3979574 T8058 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[297 (1440564869073993728)]} 0 0
   [junit4]   2> 3979575 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_1.fnm&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979577 T8023 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_0_Lucene41_0.tip&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979576 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[297 (1440564869073993728)]} 0 18
   [junit4]   2> 3979582 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_0.fnm&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979585 T8025 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_0.si&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979589 T8028 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_1_Lucene41_0.tim&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979590 T8023 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_0_Lucene41_0.tim&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979596 T8107 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_2_1.del&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979599 T8025 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_2.fdt&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979601 T8023 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_0_Lucene41_0.tip&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979606 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_0.fnm&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979608 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_1_Lucene41_0.tip&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979615 T8005 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[242 (-1440564869131665408)]} 0 3
   [junit4]   2> 3979618 T8025 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_1_Lucene41_0.tim&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979627 T8028 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_2_Lucene41_0.doc&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979629 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3979632 T8107 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_2.fdt&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979637 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_1_Lucene41_0.tip&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979637 T8041 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[242 (-1440564869148442624)]} 0 0
   [junit4]   2> 3979639 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_1_1.del&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979640 T8025 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[242 (-1440564869148442624)]} 0 6
   [junit4]   2> 3979646 T8025 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_2.fdx&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979646 T8028 C1863 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[242 (-1440564869148442624)]} 0 1
   [junit4]   2> 3979648 T8107 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_2_Lucene41_0.doc&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979654 T8056 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[242 (-1440564869148442624)]} 0 0
   [junit4]   2> 3979656 T8025 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_2.fnm&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979658 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_2.fdx&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979658 T8023 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[242 (-1440564869148442624)]} 0 29
   [junit4]   2> 3979661 T8039 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[242]} 0 41
   [junit4]   2> 3979663 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_1.fdx&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979667 T8006 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[298 (1440564869187239936)]} 0 0
   [junit4]   2> 3979668 T8025 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_2.fnm&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979671 T8025 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_1.fdt&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979674 T8025 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 298
   [junit4]   2> 3979676 T8107 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_1.fdx&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979678 T8028 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_0_Lucene41_0.doc&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979683 T8042 C1868 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[298 (1440564869195628544)]} 0 0
   [junit4]   2> 3979685 T8028 C1862 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[298 (1440564869195628544)]} 0 5
   [junit4]   2> 3979687 T8028 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_1.si&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979687 T8023 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_1_2.del&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979689 T8057 C1864 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[298 (1440564869195628544)]} 0 0
   [junit4]   2> 3979691 T8025 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[298 (1440564869195628544)]} 0 18
   [junit4]   2> 3979692 T8043 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[298]} 0 22
   [junit4]   2> 3979696 T8107 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_0_Lucene41_0.doc&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979698 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_0.fdx&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979704 T8025 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_0_2.del&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979706 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_1.fdt&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979711 T8107 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=_0.fdt&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979713 T8107 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_1.si&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979717 T8025 C1862 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/replication params={file=segments_3&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979719 T8025 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_0.fdx&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979719 T8103 C1868 P51258 oash.SnapPuller.fetchLatestIndex Total time taken for download : 0 secs
   [junit4]   2> 3979722 T8103 C1868 P51258 oash.SnapPuller.modifyIndexProps New index installed. Updating index properties... index=index.20130714202156647
   [junit4]   2> 3979723 T8103 C1868 P51258 oash.SnapPuller.fetchLatestIndex removing old index directory MockDirWrapper(org.apache.lucene.store.RAMDirectory@356d4ac1 lockFactory=org.apache.lucene.store.NativeFSLockFactory@3033ac10)
   [junit4]   2> 3979723 T8103 C1868 P51258 oasu.DefaultSolrCoreState.newIndexWriter Creating new IndexWriter...
   [junit4]   2> 3979724 T8103 C1868 P51258 oasu.DefaultSolrCoreState.newIndexWriter Waiting until IndexWriter is unused... core=collection1_shard1_0_replica2
   [junit4]   2> 3979724 T8103 C1868 P51258 oasu.DefaultSolrCoreState.newIndexWriter Rollback old IndexWriter... core=collection1_shard1_0_replica2
   [junit4]   2> 3979725 T8004 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[158 (-1440564869249105920)]} 0 0
   [junit4]   2> 3979728 T8023 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_0_2.del&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979734 T8103 C1868 P51258 oasc.SolrCore.getNewIndexDir New index directory detected: old=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index/ new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index.20130714202156647
   [junit4]   2> 3979739 T8028 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=_0.fdt&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=1 
   [junit4]   2> 3979739 T8103 C1868 P51258 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2f254955 lockFactory=org.apache.lucene.store.NativeFSLockFactory@23e45ac7),segFN=segments_3,generation=3}
   [junit4]   2> 3979741 T8103 C1868 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3979741 T8072 C1866 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[158 (-1440564869262737408)]} 0 1
   [junit4]   2> 3979743 T8040 C1865 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[158 (-1440564869262737408)]} 0 5
   [junit4]   2> 3979744 T8103 C1868 P51258 oasu.DefaultSolrCoreState.newIndexWriter New IndexWriter is ready to be used.
   [junit4]   2> 3979745 T8107 C1861 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[158]} 0 10
   [junit4]   2> 3979750 T8003 C1860 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[299 (1440564869274271744)]} 0 1
   [junit4]   2> 3979751 T8025 C1863 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/replication params={file=segments_3&command=filecontent&checksum=true&wt=filestream&qt=/replication&generation=3} status=0 QTime=0 
   [junit4]   2> 3979751 T8103 C1868 P51258 oass.SolrIndexSearcher.<init> Opening Searcher@16d77b13 main
   [junit4]   2> 3979754 T8102 oasc.SolrCore.registerSearcher [collection1_shard1_0_replica2] Registered new searcher Searcher@16d77b13 main{StandardDirectoryReader(segments_3:9:nrt _0(5.0):C26/9 _1(5.0):C8/4 _2(5.0):C25/4)}
   [junit4]   2> 3979755 T8103 C1868 P51258 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index;done=true>>]
   [junit4]   2> 3979755 T8103 C1868 P51258 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index
   [junit4]   2> 3979754 T8023 C1861 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 299
   [junit4]   2> 3979757 T8103 C1868 P51258 oasc.RecoveryStrategy.replay Replaying buffered documents. core=collection1_shard1_0_replica2
   [junit4]   2> 3979757 T8112 C1868 P51258 oasu.UpdateLog$LogReplayer.doReplay WARN Starting log replay tlog{file=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140/collection1_shard1_0_replica2/data/tlog/tlog.0000000000000000000 refcount=2} active=true starting pos=835
   [junit4]   2> 3979755 T8105 C1867 P51251 oash.SnapPuller.fetchLatestIndex Total time taken for download : 0 secs
   [junit4]   2> 3979758 T8105 C1867 P51251 oash.SnapPuller.modifyIndexProps New index installed. Updating index properties... index=index.20130714202156725
   [junit4]   2> ASYNC  NEW_CORE C1869 name=collection1_shard1_0_replica2 org.apache.solr.core.SolrCore@200e6d1f url=http://127.0.0.1:51258/qb/collection1_shard1_0_replica2 node=127.0.0.1:51258_qb C1869_STATE=coll:collection1 core:collection1_shard1_0_replica2 props:{state=recovering, core=collection1_shard1_0_replica2, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb}
   [junit4]   2> 3979776 T8112 C1869 P51258 oasu.DirectUpdateHandler2.commit start commit{flags=2,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> ASYNC  NEW_CORE C1870 name=collection1_shard1_1_replica2 org.apache.solr.core.SolrCore@1562478b url=http://127.0.0.1:51251/qb/collection1_shard1_1_replica2 node=127.0.0.1:51251_qb C1870_STATE=coll:collection1 core:collection1_shard1_1_replica2 props:{state=recovering, core=collection1_shard1_1_replica2, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb}
   [junit4]   2> 3979779 T8105 C1870 P51251 oash.SnapPuller.fetchLatestIndex removing old index directory MockDirWrapper(org.apache.lucene.store.RAMDirectory@7db675f5 lockFactory=org.apache.lucene.store.NativeFSLockFactory@380ad5c4)
   [junit4]   2> 3979790 T8112 C1869 P51258 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2f254955 lockFactory=org.apache.lucene.store.NativeFSLockFactory@23e45ac7),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2f254955 lockFactory=org.apache.lucene.store.NativeFSLockFactory@23e45ac7),segFN=segments_4,generation=4}
   [junit4]   2> 3979796 T8112 C1869 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 4
   [junit4]   2> 3979781 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[299 (1440564869279514624)]} 0 1
   [junit4]   2> 3979797 T8105 C1870 P51251 oasu.DefaultSolrCoreState.newIndexWriter Creating new IndexWriter...
   [junit4]   2> 3979807 T8105 C1870 P51251 oasu.DefaultSolrCoreState.newIndexWriter Waiting until IndexWriter is unused... core=collection1_shard1_1_replica2
   [junit4]   2> 3979807 T8105 C1870 P51251 oasu.DefaultSolrCoreState.newIndexWriter Rollback old IndexWriter... core=collection1_shard1_1_replica2
   [junit4]   2> ASYNC  NEW_CORE C1871 name=collection1_shard1_1_replica1 org.apache.solr.core.SolrCore@54aa02ed url=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1 node=127.0.0.1:51255_qb C1871_STATE=coll:collection1 core:collection1_shard1_1_replica1 props:{state=active, core=collection1_shard1_1_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3979808 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[299 (1440564869279514624)]} 0 32
   [junit4]   2> 3979818 T8112 C1869 P51258 oass.SolrIndexSearcher.<init> Opening Searcher@34bcd60a main
   [junit4]   2> 3979819 T8112 C1869 P51258 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3979820 T8102 oasc.SolrCore.registerSearcher [collection1_shard1_0_replica2] Registered new searcher Searcher@34bcd60a main{StandardDirectoryReader(segments_4:12:nrt _0(5.0):C26/9 _1(5.0):C8/4 _2(5.0):C25/5 _3(5.0):C2)}
   [junit4]   2> ASYNC  NEW_CORE C1872 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1872_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb}
   [junit4]   2> 3979820 T8053 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[299 (1440564869279514624)]} 0 1
   [junit4]   2> 3979822 T8112 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] {add=[297 (1440564869073993728), 298 (1440564869195628544)],delete=[242 (-1440564869148442624)]} 0 65
   [junit4]   2> 3979822 T8112 C1869 P51258 oasu.UpdateLog$LogReplayer.run WARN Log replay finished. recoveryInfo=RecoveryInfo{adds=2 deletes=1 deleteByQuery=0 errors=0 positionOfStart=835}
   [junit4]   2> ASYNC  NEW_CORE C1873 name=collection1 org.apache.solr.core.SolrCore@31a4adfc url=http://127.0.0.1:51255/qb/collection1 node=127.0.0.1:51255_qb C1873_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3979822 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[299 (1440564869279514624)]} 0 68
   [junit4]   2> 3979832 T8103 C1869 P51258 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=collection1_shard1_0_replica2
   [junit4]   2> 3979835 T8103 C1869 P51258 oasc.ZkController.publish publishing core=collection1_shard1_0_replica2 state=active
   [junit4]   2> 3979835 T8103 C1869 P51258 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3979837 T8103 C1869 P51258 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=collection1_shard1_0_replica2
   [junit4]   2> 3979838 T8105 C1870 P51251 oasc.SolrCore.getNewIndexDir New index directory detected: old=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index/ new=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index.20130714202156725
   [junit4]   2> 3979842 T8105 C1870 P51251 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@22ec99d4 lockFactory=org.apache.lucene.store.NativeFSLockFactory@22d86580),segFN=segments_3,generation=3}
   [junit4]   2> 3979842 T8105 C1870 P51251 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3979843 T8105 C1870 P51251 oasu.DefaultSolrCoreState.newIndexWriter New IndexWriter is ready to be used.
   [junit4]   2> 3979847 T8105 C1870 P51251 oass.SolrIndexSearcher.<init> Opening Searcher@713173f7 main
   [junit4]   2> 3979849 T8104 oasc.SolrCore.registerSearcher [collection1_shard1_1_replica2] Registered new searcher Searcher@713173f7 main{StandardDirectoryReader(segments_3:9:nrt _0(5.0):C54/9 _1(5.0):C8/4 _2(5.0):C29/7)}
   [junit4]   2> 3979849 T8105 C1870 P51251 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index;done=true>>]
   [junit4]   2> 3979850 T8105 C1870 P51251 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index
   [junit4]   2> 3979851 T8105 C1870 P51251 oasc.RecoveryStrategy.replay Replaying buffered documents. core=collection1_shard1_1_replica2
   [junit4]   2> 3979851 T8113 C1870 P51251 oasu.UpdateLog$LogReplayer.doReplay WARN Starting log replay tlog{file=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446/collection1_shard1_1_replica2/data/tlog/tlog.0000000000000000000 refcount=2} active=true starting pos=651
   [junit4]   2> 3979853 T8113 C1870 P51251 oasu.DirectUpdateHandler2.commit start commit{flags=2,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3979856 T8113 C1870 P51251 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@22ec99d4 lockFactory=org.apache.lucene.store.NativeFSLockFactory@22d86580),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@22ec99d4 lockFactory=org.apache.lucene.store.NativeFSLockFactory@22d86580),segFN=segments_4,generation=4}
   [junit4]   2> 3979857 T8113 C1870 P51251 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 4
   [junit4]   2> 3979859 T8113 C1870 P51251 oass.SolrIndexSearcher.<init> Opening Searcher@7b0d0647 main
   [junit4]   2> 3979860 T8113 C1870 P51251 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3979861 T8104 oasc.SolrCore.registerSearcher [collection1_shard1_1_replica2] Registered new searcher Searcher@7b0d0647 main{StandardDirectoryReader(segments_4:11:nrt _0(5.0):C54/9 _1(5.0):C8/4 _2(5.0):C29/7 _3(5.0):C2)}
   [junit4]   2> 3979862 T8113 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] {add=[296 (1440564868979621888), 299 (1440564869279514624)]} 0 11
   [junit4]   2> 3979862 T8113 C1870 P51251 oasu.UpdateLog$LogReplayer.run WARN Log replay finished. recoveryInfo=RecoveryInfo{adds=2 deletes=0 deleteByQuery=0 errors=0 positionOfStart=651}
   [junit4]   2> 3979863 T8105 C1870 P51251 oasc.RecoveryStrategy.doRecovery Replication Recovery was successful - registering as Active. core=collection1_shard1_1_replica2
   [junit4]   2> 3979863 T8105 C1870 P51251 oasc.ZkController.publish publishing core=collection1_shard1_1_replica2 state=active
   [junit4]   2> 3979864 T8105 C1870 P51251 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> ASYNC  NEW_CORE C1874 name=collection1 org.apache.solr.core.SolrCore@7a780607 url=http://127.0.0.1:51251/qb/collection1 node=127.0.0.1:51251_qb C1874_STATE=coll:control_collection core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3979867 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[284 (-1440564869398003712)]} 0 0
   [junit4]   2> 3979871 T8105 C1870 P51251 oasc.RecoveryStrategy.doRecovery Finished recovery process. core=collection1_shard1_1_replica2
   [junit4]   2> 3979886 T8107 C1873 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> ASYNC  NEW_CORE C1875 name=collection1_shard1_0_replica1 org.apache.solr.core.SolrCore@5c1f6bce url=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1 node=127.0.0.1:51255_qb C1875_STATE=coll:collection1 core:collection1_shard1_0_replica1 props:{state=active, core=collection1_shard1_0_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3979898 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[284 (-1440564869417926656)]} 0 2
   [junit4]   2> 3979909 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[284 (-1440564869417926656)]} 0 2
   [junit4]   2> 3979917 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[284 (-1440564869417926656)]} 0 14
   [junit4]   2> 3979922 T8055 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[284 (-1440564869417926656)]} 0 0
   [junit4]   2> 3979924 T8107 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[284 (-1440564869417926656)]} 0 38
   [junit4]   2> ASYNC  NEW_CORE C1876 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1876_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 3979925 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[284]} 0 52
   [junit4]   2> 3979940 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[300 (1440564869474549760)]} 0 0
   [junit4]   2> 3979947 T8025 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 300
   [junit4]   2> 3979953 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[300 (1440564869481889792)]} 0 1
   [junit4]   2> 3979954 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[300 (1440564869481889792)]} 0 4
   [junit4]   2> 3979958 T8059 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[300 (1440564869481889792)]} 0 0
   [junit4]   2> 3979959 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[300 (1440564869481889792)]} 0 13
   [junit4]   2> 3979960 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[300]} 0 16
   [junit4]   2> 3979994 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[301 (1440564869530124288)]} 0 1
   [junit4]   2> 3980008 T8023 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 301
   [junit4]   2> 3980019 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[301 (1440564869544804352)]} 0 1
   [junit4]   2> 3980021 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[301 (1440564869544804352)]} 0 6
   [junit4]   2> 3980026 T8058 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[301 (1440564869544804352)]} 0 0
   [junit4]   2> 3980028 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[301 (1440564869544804352)]} 0 22
   [junit4]   2> 3980029 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[301]} 0 28
   [junit4]   2> 3980062 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[302 (1440564869602476032)]} 0 1
   [junit4]   2> ASYNC  NEW_CORE C1877 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1877_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb}
   [junit4]   2> 3980075 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[302 (1440564869612961792)]} 0 1
   [junit4]   2> 3980077 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[302 (1440564869612961792)]} 0 5
   [junit4]   2> 3980079 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[302]} 0 11
   [junit4]   2> 3980113 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[234 (-1440564869655953408)]} 0 0
   [junit4]   2> 3980118 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[234 (-1440564869659099136)]} 0 0
   [junit4]   2> 3980119 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[234 (-1440564869659099136)]} 0 3
   [junit4]   2> 3980122 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[303 (1440564869665390592)]} 0 0
   [junit4]   2> 3980126 T8025 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 303
   [junit4]   2> 3980134 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[303 (1440564869669584896)]} 0 3
   [junit4]   2> 3980135 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[303 (1440564869669584896)]} 0 6
   [junit4]   2> 3980139 T8057 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[303 (1440564869669584896)]} 0 1
   [junit4]   2> 3980140 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[303 (1440564869669584896)]} 0 15
   [junit4]   2> 3980174 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[304 (1440564869718867968)]} 0 1
   [junit4]   2> 3980183 T8023 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 304
   [junit4]   2> 3980196 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[304 (1440564869728305152)]} 0 1
   [junit4]   2> 3980197 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[304 (1440564869728305152)]} 0 10
   [junit4]   2> 3980203 T8053 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[304 (1440564869728305152)]} 0 1
   [junit4]   2> 3980204 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[304 (1440564869728305152)]} 0 22
   [junit4]   2> 3980240 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[272 (-1440564869789122560)]} 0 0
   [junit4]   2> 3980253 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[272 (-1440564869798559744)]} 0 0
   [junit4]   2> 3980254 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[272 (-1440564869798559744)]} 0 5
   [junit4]   2> 3980256 T8107 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[272]} 0 11
   [junit4]   2> 3980261 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[305 (1440564869811142656)]} 0 1
   [junit4]   2> 3980268 T8025 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 305
   [junit4]   2> 3980274 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[305 (1440564869817434112)]} 0 1
   [junit4]   2> 3980275 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[305 (1440564869817434112)]} 0 5
   [junit4]   2> 3980279 T8055 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[305 (1440564869817434112)]} 0 1
   [junit4]   2> 3980280 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[305 (1440564869817434112)]} 0 13
   [junit4]   2> 3980314 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[306 (1440564869865668608)]} 0 1
   [junit4]   2> 3980325 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[306 (1440564869875105792)]} 0 1
   [junit4]   2> 3980326 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[306 (1440564869875105792)]} 0 5
   [junit4]   2> 3980327 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[306]} 0 8
   [junit4]   2> 3980360 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[215 (-1440564869914951680)]} 0 0
   [junit4]   2> 3980364 T8028 C1873 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3980369 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[215 (-1440564869919145984)]} 0 0
   [junit4]   2> 3980371 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[215 (-1440564869919145984)]} 0 5
   [junit4]   2> 3980374 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[215 (-1440564869919145984)]} 0 1
   [junit4]   2> 3980383 T8059 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[215 (-1440564869919145984)]} 0 4
   [junit4]   2> 3980386 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[215 (-1440564869919145984)]} 0 23
   [junit4]   2> 3980390 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[307 (1440564869945360384)]} 0 1
   [junit4]   2> 3980399 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[307 (1440564869952700416)]} 0 1
   [junit4]   2> 3980402 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[307 (1440564869952700416)]} 0 6
   [junit4]   2> 3980403 T8107 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[307]} 0 10
   [junit4]   2> 3980442 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[229 (-1440564870000934912)]} 0 0
   [junit4]   2> 3980460 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[229 (-1440564870016663552)]} 0 1
   [junit4]   2> 3980461 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[229 (-1440564870016663552)]} 0 4
   [junit4]   2> 3980465 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[229]} 0 11
   [junit4]   2> 3980469 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[308 (1440564870028197888)]} 0 1
   [junit4]   2> 3980475 T8023 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 308
   [junit4]   2> 3980482 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[308 (1440564870035537920)]} 0 1
   [junit4]   2> 3980486 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[308 (1440564870035537920)]} 0 8
   [junit4]   2> 3980490 T8056 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[308 (1440564870035537920)]} 0 1
   [junit4]   2> 3980491 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[308 (1440564870035537920)]} 0 17
   [junit4]   2> 3980492 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[308]} 0 20
   [junit4]   2> 3980531 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[309 (1440564870093209600)]} 0 1
   [junit4]   2> 3980537 T8107 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 309
   [junit4]   2> 3980544 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[309 (1440564870100549632)]} 0 0
   [junit4]   2> 3980548 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[309 (1440564870100549632)]} 0 8
   [junit4]   2> 3980563 T8057 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[309 (1440564870100549632)]} 0 2
   [junit4]   2> 3980564 T8107 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[309 (1440564870100549632)]} 0 27
   [junit4]   2> 3980565 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[309]} 0 31
   [junit4]   2> 3980599 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[310 (1440564870164512768)]} 0 1
   [junit4]   2> 3980603 T8028 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 310
   [junit4]   2> 3980613 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[310 (1440564870169755648)]} 0 1
   [junit4]   2> 3980615 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[310 (1440564870169755648)]} 0 7
   [junit4]   2> 3980620 T8053 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[310 (1440564870169755648)]} 0 0
   [junit4]   2> 3980622 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[310 (1440564870169755648)]} 0 19
   [junit4]   2> 3980657 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[208 (-1440564870225330176)]} 0 2
   [junit4]   2> 3980673 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[208 (-1440564870237913088)]} 0 1
   [junit4]   2> 3980674 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[208 (-1440564870237913088)]} 0 6
   [junit4]   2> 3980675 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[208]} 0 11
   [junit4]   2> 3980680 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[311 (1440564870249447424)]} 0 1
   [junit4]   2> 3980688 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[311 (1440564870254690304)]} 0 1
   [junit4]   2> 3980689 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[311 (1440564870254690304)]} 0 6
   [junit4]   2> 3980727 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[285 (-1440564870299779072)]} 0 1
   [junit4]   2> 3980735 T8107 C1873 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3980739 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[285 (-1440564870308167680)]} 0 1
   [junit4]   2> 3980746 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[285 (-1440564870308167680)]} 0 0
   [junit4]   2> 3980749 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[285 (-1440564870308167680)]} 0 7
   [junit4]   2> 3980753 T8055 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[285 (-1440564870308167680)]} 0 0
   [junit4]   2> 3980754 T8107 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[285 (-1440564870308167680)]} 0 19
   [junit4]   2> 3980769 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[285]} 0 38
   [junit4]   2> 3980773 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[312 (1440564870348013568)]} 0 1
   [junit4]   2> 3980783 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[312 (1440564870353256448)]} 0 1
   [junit4]   2> 3980784 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[312 (1440564870353256448)]} 0 6
   [junit4]   2> 3980832 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[313 (1440564870408830976)]} 0 1
   [junit4]   2> 3980850 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[313 (1440564870424559616)]} 0 1
   [junit4]   2> 3980851 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[313 (1440564870424559616)]} 0 5
   [junit4]   2> 3980852 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[313]} 0 9
   [junit4]   2> 3980886 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[314 (1440564870465454080)]} 0 1
   [junit4]   2> 3980892 T8028 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 314
   [junit4]   2> 3980898 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[314 (1440564870471745536)]} 0 1
   [junit4]   2> 3980899 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[314 (1440564870471745536)]} 0 4
   [junit4]   2> 3980904 T8059 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[314 (1440564870471745536)]} 0 1
   [junit4]   2> 3980905 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[314 (1440564870471745536)]} 0 14
   [junit4]   2> 3980906 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[314]} 0 17
   [junit4]   2> 3980941 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[227 (-1440564870524174336)]} 0 1
   [junit4]   2> 3980945 T8107 C1873 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3980952 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[227 (-1440564870528368640)]} 0 1
   [junit4]   2> 3980954 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[227 (-1440564870528368640)]} 0 6
   [junit4]   2> 3980958 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[227 (-1440564870528368640)]} 0 1
   [junit4]   2> 3980962 T8056 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[227 (-1440564870528368640)]} 0 1
   [junit4]   2> 3980963 T8107 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[227 (-1440564870528368640)]} 0 18
   [junit4]   2> 3980966 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[315 (1440564870550388736)]} 0 0
   [junit4]   2> 3980973 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[315 (1440564870554583040)]} 0 1
   [junit4]   2> 3980974 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[315 (1440564870554583040)]} 0 5
   [junit4]   2> 3981009 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[261 (-1440564870594428928)]} 0 1
   [junit4]   2> 3981017 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[261 (-1440564870600720384)]} 0 1
   [junit4]   2> 3981018 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[261 (-1440564870600720384)]} 0 4
   [junit4]   2> 3981022 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[316 (1440564870609108992)]} 0 1
   [junit4]   2> 3981026 T8028 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 316
   [junit4]   2> 3981032 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[316 (1440564870612254720)]} 0 1
   [junit4]   2> 3981034 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[316 (1440564870612254720)]} 0 5
   [junit4]   2> 3981039 T8057 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[316 (1440564870612254720)]} 0 1
   [junit4]   2> 3981040 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[316 (1440564870612254720)]} 0 15
   [junit4]   2> 3981073 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[317 (1440564870662586368)]} 0 1
   [junit4]   2> 3981082 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[317 (1440564870667829248)]} 0 1
   [junit4]   2> 3981084 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[317 (1440564870667829248)]} 0 7
   [junit4]   2> 3981122 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[318 (1440564870712918016)]} 0 1
   [junit4]   2> 3981129 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[318 (1440564870718160896)]} 0 1
   [junit4]   2> 3981130 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[318 (1440564870718160896)]} 0 4
   [junit4]   2> 3981156 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3981160 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_0",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1_shard1_0_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 3981164 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[319 (1440564870756958208)]} 0 1
   [junit4]   2> 3981171 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[319 (1440564870762201088)]} 0 0
   [junit4]   2> 3981173 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[319 (1440564870762201088)]} 0 5
   [junit4]   2> 3981178 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1_shard1_1_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3981193 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3981193 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3981194 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3981193 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3981195 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3981193 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3981208 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[320 (1440564870803095552)]} 0 2
   [junit4]   2> 3981215 T8025 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 320
   [junit4]   2>  C1870_STATE=coll:collection1 core:collection1_shard1_1_replica2 props:{state=active, core=collection1_shard1_1_replica2, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb}
   [junit4]   2> 3981222 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[320 (1440564870811484160)]} 0 1
   [junit4]   2> 3981224 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[320 (1440564870811484160)]} 0 5
   [junit4]   2> 3981227 T8053 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[320 (1440564870811484160)]} 0 0
   [junit4]   2> 3981229 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[320 (1440564870811484160)]} 0 14
   [junit4]   2> 3981230 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[320]} 0 18
   [junit4]   2> 3981263 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[166 (-1440564870861815808)]} 0 0
   [junit4]   2> 3981271 T8028 C1873 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2>  C1869_STATE=coll:collection1 core:collection1_shard1_0_replica2 props:{state=active, core=collection1_shard1_0_replica2, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb}
   [junit4]   2> 3981278 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[166 (-1440564870869155840)]} 0 0
   [junit4]   2> 3981280 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[166 (-1440564870869155840)]} 0 5
   [junit4]   2> 3981284 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[166 (-1440564870869155840)]} 0 0
   [junit4]   2> 3981288 T8055 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[166 (-1440564870869155840)]} 0 0
   [junit4]   2> 3981290 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[166 (-1440564870869155840)]} 0 20
   [junit4]   2> 3981292 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[166]} 0 24
   [junit4]   2> 3981296 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[321 (1440564870895370240)]} 0 1
   [junit4]   2> 3981300 T8025 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 321
   [junit4]   2> 3981307 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[321 (1440564870900613120)]} 0 1
   [junit4]   2> 3981308 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[321 (1440564870900613120)]} 0 5
   [junit4]   2> 3981313 T8058 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[321 (1440564870900613120)]} 0 1
   [junit4]   2> 3981314 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[321 (1440564870900613120)]} 0 15
   [junit4]   2> 3981348 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[322 (1440564870949896192)]} 0 1
   [junit4]   2> 3981360 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[322 (1440564870957236224)]} 0 1
   [junit4]   2> 3981361 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[322 (1440564870957236224)]} 0 7
   [junit4]   2> 3981363 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[322]} 0 12
   [junit4]   2> 3981400 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[323 (1440564871004422144)]} 0 1
   [junit4]   2> 3981406 T8028 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 323
   [junit4]   2> 3981412 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[323 (1440564871010713600)]} 0 1
   [junit4]   2> 3981413 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[323 (1440564871010713600)]} 0 5
   [junit4]   2> 3981418 T8056 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[323 (1440564871010713600)]} 0 1
   [junit4]   2> 3981419 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[323 (1440564871010713600)]} 0 14
   [junit4]   2> 3981455 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[324 (1440564871063142400)]} 0 1
   [junit4]   2> 3981473 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[324 (1440564871075725312)]} 0 0
   [junit4]   2> 3981475 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[324 (1440564871075725312)]} 0 8
   [junit4]   2> 3981476 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[324]} 0 15
   [junit4]   2> 3981513 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[325 (1440564871118716928)]} 0 5
   [junit4]   2> 3981518 T8023 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 325
   [junit4]   2> 3981528 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[325 (1440564871129202688)]} 0 1
   [junit4]   2> 3981529 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[325 (1440564871129202688)]} 0 5
   [junit4]   2> 3981533 T8057 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[325 (1440564871129202688)]} 0 0
   [junit4]   2> 3981534 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[325 (1440564871129202688)]} 0 17
   [junit4]   2> 3981568 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[326 (1440564871180582912)]} 0 1
   [junit4]   2> 3981581 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[326 (1440564871187922944)]} 0 1
   [junit4]   2> 3981583 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[326 (1440564871187922944)]} 0 9
   [junit4]   2> 3981584 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[326]} 0 13
   [junit4]   2> 3981618 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[260 (-1440564871234060288)]} 0 0
   [junit4]   2> 3981626 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[260 (-1440564871240351744)]} 0 0
   [junit4]   2> 3981628 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[260 (-1440564871240351744)]} 0 4
   [junit4]   2> 3981629 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[260]} 0 8
   [junit4]   2> 3981633 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[327 (1440564871248740352)]} 0 1
   [junit4]   2> 3981650 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[327 (1440564871255031808)]} 0 1
   [junit4]   2> 3981652 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[327 (1440564871255031808)]} 0 14
   [junit4]   2> 3981653 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[327]} 0 17
   [junit4]   2> 3981687 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[328 (1440564871305363456)]} 0 1
   [junit4]   2> 3981694 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[328 (1440564871310606336)]} 0 0
   [junit4]   2> 3981696 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[328 (1440564871310606336)]} 0 5
   [junit4]   2> 3981729 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[147 (-1440564871350452224)]} 0 1
   [junit4]   2> 3981736 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[147 (-1440564871354646528)]} 0 0
   [junit4]   2> 3981737 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[147 (-1440564871354646528)]} 0 4
   [junit4]   2> 3981741 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[329 (1440564871361986560)]} 0 1
   [junit4]   2> 3981747 T8028 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 329
   [junit4]   2> 3981754 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[329 (1440564871369326592)]} 0 1
   [junit4]   2> 3981755 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[329 (1440564871369326592)]} 0 4
   [junit4]   2> 3981758 T8053 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[329 (1440564871369326592)]} 0 0
   [junit4]   2> 3981759 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[329 (1440564871369326592)]} 0 13
   [junit4]   2> 3981760 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[329]} 0 16
   [junit4]   2> 3981794 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[223 (-1440564871418609664)]} 0 1
   [junit4]   2> 3981802 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[223 (-1440564871423852544)]} 0 1
   [junit4]   2> 3981803 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[223 (-1440564871423852544)]} 0 5
   [junit4]   2> 3981807 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[330 (1440564871432241152)]} 0 1
   [junit4]   2> 3981815 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[330 (1440564871436435456)]} 0 1
   [junit4]   2> 3981817 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[330 (1440564871436435456)]} 0 7
   [junit4]   2> 3981827 T8026 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node7, state: active, checkLive: true, onlyIfLeader: true for: 9 seconds.
   [junit4]   2> 3981827 T8026 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node7&state=active&nodeName=127.0.0.1:51258_qb&action=PREPRECOVERY&checkLive=true&core=collection1_shard1_0_replica1&wt=javabin&qt=/admin/cores&onlyIfLeader=true&version=2} status=0 QTime=9016 
   [junit4]   2> 3981831 T8024 oasha.CoreAdminHandler.handleWaitForStateAction Waited coreNodeName: core_node8, state: active, checkLive: true, onlyIfLeader: true for: 8 seconds.
   [junit4]   2> 3981832 T8024 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/cores params={coreNodeName=core_node8&state=active&nodeName=127.0.0.1:51251_qb&action=PREPRECOVERY&checkLive=true&core=collection1_shard1_1_replica1&wt=javabin&qt=/admin/cores&onlyIfLeader=true&version=2} status=0 QTime=8019 
   [junit4]   2> 3981833 T8014 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3981855 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[331 (1440564871481524224)]} 0 1
   [junit4]   2> 3981859 T8025 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 331
   [junit4]   2> 3981863 T8107 C1873 P51255 oasu.SolrCmdDistributor.distribCommit Distrib commit to:[StdNode: http://127.0.0.1:51255/qb/collection1/, StdNode: http://127.0.0.1:51261/qb/collection1/, StdNode: http://127.0.0.1:51258/qb/collection1/, StdNode: http://127.0.0.1:51265/qb/collection1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51258/qb/collection1_shard1_0_replica2/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/, StdNode: http://127.0.0.1:51251/qb/collection1_shard1_1_replica2/] params:commit_end_point=true&commit=true&softCommit=false&waitSearcher=true&expungeDeletes=false
   [junit4]   2> 3981865 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[331 (1440564871486767104)]} 0 0
   [junit4]   2> 3981867 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[331 (1440564871486767104)]} 0 5
   [junit4]   2> 3981868 T8023 C1873 P51255 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3981871 T8059 C1872 P51261 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3981878 T8041 C1876 P51258 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3981884 T8026 C1875 P51255 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3981886 T8042 C1869 P51258 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3981898 T8073 C1877 P51265 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3981917 T8003 C1870 P51251 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3981920 T8024 C1871 P51255 oasu.DirectUpdateHandler2.commit start commit{,optimize=false,openSearcher=true,waitSearcher=true,expungeDeletes=false,softCommit=false,prepareCommit=false}
   [junit4]   2> 3981937 T8058 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[331 (1440564871486767104)]} 0 15
   [junit4]   2> 3981939 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[331 (1440564871486767104)]} 0 81
   [junit4]   2> 3981964 T8003 C1870 P51251 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@22ec99d4 lockFactory=org.apache.lucene.store.NativeFSLockFactory@22d86580),segFN=segments_4,generation=4}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@22ec99d4 lockFactory=org.apache.lucene.store.NativeFSLockFactory@22d86580),segFN=segments_5,generation=5}
   [junit4]   2> 3981965 T8003 C1870 P51251 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 5
   [junit4]   2> 3981967 T8003 C1870 P51251 oass.SolrIndexSearcher.<init> Opening Searcher@36a1c96a main
   [junit4]   2> 3981968 T8003 C1870 P51251 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3981969 T8104 oasc.SolrCore.registerSearcher [collection1_shard1_1_replica2] Registered new searcher Searcher@36a1c96a main{StandardDirectoryReader(segments_5:14:nrt _0(5.0):C54/9 _1(5.0):C8/4 _2(5.0):C29/9 _3(5.0):C2 _4(5.0):C9)}
   [junit4]   2> 3981970 T8059 C1872 P51261 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@4cf6277a lockFactory=org.apache.lucene.store.NativeFSLockFactory@518a7bdf),segFN=segments_2,generation=2}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@4cf6277a lockFactory=org.apache.lucene.store.NativeFSLockFactory@518a7bdf),segFN=segments_3,generation=3}
   [junit4]   2> 3981971 T8059 C1872 P51261 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3981970 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 53
   [junit4]   2> 3981978 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[332 (1440564871611547648)]} 0 1
   [junit4]   2> 3981979 T8023 C1873 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_4,generation=4}
   [junit4]   2> 3981979 T8023 C1873 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 4
   [junit4]   2> 3981984 T8073 C1877 P51265 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191fa213 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1b45ecd0),segFN=segments_2,generation=2}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191fa213 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1b45ecd0),segFN=segments_3,generation=3}
   [junit4]   2> 3981984 T8073 C1877 P51265 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3981997 T8041 C1876 P51258 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=3
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_1,generation=1}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_2,generation=2}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_3,generation=3}
   [junit4]   2> 3981998 T8041 C1876 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3982001 T8041 C1876 P51258 oass.SolrIndexSearcher.<init> Opening Searcher@5432e0ed main
   [junit4]   2> 3982002 T8041 C1876 P51258 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3982003 T8050 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@5432e0ed main{StandardDirectoryReader(segments_3:5:nrt _0(5.0):C42 _1(5.0):C118/62)}
   [junit4]   2> 3982004 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 127
   [junit4]   2> 3982009 T8073 C1877 P51265 oass.SolrIndexSearcher.<init> Opening Searcher@5f472ea8 main
   [junit4]   2> 3982010 T8028 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 332
   [junit4]   2> 3982011 T8073 C1877 P51265 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3982012 T8059 C1872 P51261 oass.SolrIndexSearcher.<init> Opening Searcher@38c587d4 main
   [junit4]   2> 3982014 T8084 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@5f472ea8 main{StandardDirectoryReader(segments_3:5:nrt _0(5.0):C42 _1(5.0):C118/62)}
   [junit4]   2> 3982016 T8066 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@38c587d4 main{StandardDirectoryReader(segments_3:7:nrt _0(5.0):C59 _1(5.0):C112/45 _2(5.0):C1)}
   [junit4]   2> 3982014 T8059 C1872 P51261 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3982018 T8059 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 147
   [junit4]   2> 3982018 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 120
   [junit4]   2> 3982021 T8026 C1875 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_4,generation=4}
   [junit4]   2> 3982023 T8026 C1875 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 4
   [junit4]   2> 3982021 T8024 C1871 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_4,generation=4}
   [junit4]   2> 3982031 T8024 C1871 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 4
   [junit4]   2> 3982020 T8042 C1869 P51258 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2f254955 lockFactory=org.apache.lucene.store.NativeFSLockFactory@23e45ac7),segFN=segments_4,generation=4}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2f254955 lockFactory=org.apache.lucene.store.NativeFSLockFactory@23e45ac7),segFN=segments_5,generation=5}
   [junit4]   2> 3982033 T8042 C1869 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 5
   [junit4]   2> 3982046 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[332 (1440564871631470592)]} 0 8
   [junit4]   2> 3982048 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[332 (1440564871631470592)]} 0 33
   [junit4]   2> 3982060 T8026 C1875 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@10c5570f main
   [junit4]   2> 3982062 T8026 C1875 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3982063 T8097 oasc.SolrCore.registerSearcher [collection1_shard1_0_replica1] Registered new searcher Searcher@10c5570f main{StandardDirectoryReader(segments_4:12:nrt _0(5.0):C26/9 _1(5.0):C8/5 _2(5.0):C25/7 _3(5.0):C9)}
   [junit4]   2> 3982064 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 180
   [junit4]   2> 3982067 T8042 C1869 P51258 oass.SolrIndexSearcher.<init> Opening Searcher@a04d1e5 main
   [junit4]   2> 3982068 T8042 C1869 P51258 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3982069 T8024 C1871 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@591dff6c main
   [junit4]   2> 3982071 T8102 oasc.SolrCore.registerSearcher [collection1_shard1_0_replica2] Registered new searcher Searcher@a04d1e5 main{StandardDirectoryReader(segments_5:15:nrt _0(5.0):C26/9 _1(5.0):C8/5 _2(5.0):C25/7 _3(5.0):C2 _4(5.0):C7)}
   [junit4]   2> 3982073 T8056 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[332 (1440564871631470592)]} 0 1
   [junit4]   2> 3982074 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 188
   [junit4]   2> 3982075 T8096 oasc.SolrCore.registerSearcher [collection1_shard1_1_replica1] Registered new searcher Searcher@591dff6c main{StandardDirectoryReader(segments_4:14:nrt _0(5.0):C54/9 _1(5.0):C8/4 _2(5.0):C29/9 _3(5.0):C11 _4(5.0):C1)}
   [junit4]   2> 3982077 T8024 C1871 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3982077 T8024 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 158
   [junit4]   2> 3982079 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[332 (1440564871631470592)]} 0 82
   [junit4]   2> 3982080 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[332]} 0 95
   [junit4]   2> 3982085 T8023 C1873 P51255 oass.SolrIndexSearcher.<init> Opening Searcher@37f6b4f6 main
   [junit4]   2> 3982087 T8023 C1873 P51255 oasu.DirectUpdateHandler2.commit end_commit_flush
   [junit4]   2> 3982088 T8034 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@37f6b4f6 main{StandardDirectoryReader(segments_4:10:nrt _0(5.0):C59 _1(5.0):C39/29 _2(5.0):C74/16 _3(5.0):C1)}
   [junit4]   2> 3982096 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&commit=true&wt=javabin&expungeDeletes=false&commit_end_point=true&version=2&softCommit=false} {commit=} 0 228
   [junit4]   2> 3982098 T8107 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={waitSearcher=true&openSearcher=true&commit=true&wt=javabin&version=2&softCommit=false} {commit=} 0 235
   [junit4]   2> 3982109 T8014 oasc.OverseerCollectionProcessor.splitShard Successfully created all replica shards for all sub-slices [shard1_0, shard1_1]
   [junit4]   2> 3982110 T8014 oasc.OverseerCollectionProcessor.splitShard Requesting update shard state
   [junit4]   2> 3982122 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[241 (-1440564871761494016)]} 0 1
   [junit4]   2> 3982139 T8032 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work/qnr-0000000000 state: SyncConnected type NodeDataChanged
   [junit4]   2> 3982148 T8027 oass.SolrDispatchFilter.handleAdminRequest [admin] webapp=null path=/admin/collections params={shard=shard1&action=SPLITSHARD&collection=collection1&wt=javabin&version=2} status=0 QTime=17533 
   [junit4]   2> 3982151 T7990 oasc.ShardSplitTest.doTest Layout after split: 
   [junit4]   2> 	
   [junit4]   2> 3982153 T8026 C1873 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3982158 T7990 oascc.ConnectionManager.waitForConnected Waiting for client to connect to ZooKeeper
   [junit4]   2> 3982158 T8012 oasc.DistributedQueue$LatchChildWatcher.process Watcher fired on path: /overseer/collection-queue-work state: SyncConnected type NodeChildrenChanged
   [junit4]   2> 3982167 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[241 (-1440564871791902720)]} 0 3
   [junit4]   2> 3982168 T8014 oasc.OverseerCollectionProcessor.run Overseer Collection Processor: Message id:/overseer/collection-queue-work/qn-0000000000 complete, response:{success={null={responseHeader={status=0,QTime=4049},core=collection1_shard1_1_replica1},null={responseHeader={status=0,QTime=4055},core=collection1_shard1_0_replica1},null={responseHeader={status=0,QTime=2004}},null={responseHeader={status=0,QTime=2005}},null={responseHeader={status=0,QTime=33}},null={responseHeader={status=0,QTime=40},core=collection1_shard1_0_replica1,status=BUFFER_APPLIED},null={responseHeader={status=0,QTime=48},core=collection1_shard1_1_replica1,status=BUFFER_APPLIED},null={responseHeader={status=0,QTime=4576},core=collection1_shard1_0_replica2},null={responseHeader={status=0,QTime=3545},core=collection1_shard1_1_replica2},null={responseHeader={status=0,QTime=9016}},null={responseHeader={status=0,QTime=8019}},http://127.0.0.1:51255/qb/collection1/={responseHeader={status=0,QTime=235}}}}
   [junit4]   2> 3982172 T8123 oascc.ConnectionManager.process Watcher org.apache.solr.common.cloud.ConnectionManager@55c44c20 name:ZooKeeperConnection Watcher:127.0.0.1:51248 got event WatchedEvent state:SyncConnected type:None path:null path:null type:None
   [junit4]   2> 3982172 T7990 oascc.ConnectionManager.waitForConnected Client is connected to ZooKeeper
   [junit4]   2> 3982182 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[241 (-1440564871791902720)]} 0 1
   [junit4]   2> 3982183 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[241 (-1440564871791902720)]} 0 5
   [junit4]   2> 3982188 T8057 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[241 (-1440564871791902720)]} 0 1
   [junit4]   2> 3982190 T8026 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[241 (-1440564871791902720)]} 0 41
   [junit4]   2> 3982191 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[241]} 0 51
   [junit4]   2> 3982198 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[333 (1440564871841185792)]} 0 1
   [junit4]   2> 3982210 T8023 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 333
   [junit4]   2> 3982227 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[333 (1440564871853768704)]} 0 1
   [junit4]   2> 3982229 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[333 (1440564871853768704)]} 0 16
   [junit4]   2> 3982247 T8053 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[333 (1440564871853768704)]} 0 1
   [junit4]   2> 3982249 T8023 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[333 (1440564871853768704)]} 0 41
   [junit4]   2> 3982250 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[333]} 0 48
   [junit4]   2> 3982287 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[334 (1440564871930314752)]} 0 5
   [junit4]   2> 3982319 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[334 (1440564871948140544)]} 0 9
   [junit4]   2> 3982326 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[334 (1440564871948140544)]} 0 27
   [junit4]   2> 3982328 T8025 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[334]} 0 32
   [junit4]   2> 3982365 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[335 (1440564872016297984)]} 0 1
   [junit4]   2> 3982375 T8027 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ : 335
   [junit4]   2> 3982405 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[335 (1440564872026783744)]} 0 3
   [junit4]   2> 3982407 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[335 (1440564872026783744)]} 0 28
   [junit4]   2> 3982412 T8058 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[335 (1440564872026783744)]} 0 1
   [junit4]   2> 3982413 T8027 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[335 (1440564872026783744)]} 0 39
   [junit4]   2> 3982414 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[335]} 0 43
   [junit4]   2> 3982458 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[336 (1440564872113815552)]} 0 1
   [junit4]   2> 3982479 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[336 (1440564872132689920)]} 0 1
   [junit4]   2> 3982481 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[336 (1440564872132689920)]} 0 6
   [junit4]   2> 3982487 T8028 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[336]} 0 24
   [junit4]   2> 3982527 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[329 (-1440564872187215872)]} 0 0
   [junit4]   2> 3982533 T8107 C1873 P51255 oasu.SolrCmdDistributor.syncDelete SYNCDELETE on [StdNode: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/, StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/] : 
   [junit4]   2> 3982536 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[329 (-1440564872192458752)]} 0 1
   [junit4]   2> 3982555 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[329 (-1440564872192458752)]} 0 1
   [junit4]   2> 3982556 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {delete=[329 (-1440564872192458752)]} 0 5
   [junit4]   2> 3982562 T8059 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[329 (-1440564872192458752)]} 0 0
   [junit4]   2> 3982565 T8107 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[329 (-1440564872192458752)]} 0 33
   [junit4]   2> 3982571 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[337 (1440564872232304640)]} 0 1
   [junit4]   2> 3982601 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[337 (1440564872257470464)]} 0 1
   [junit4]   2> 3982603 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[337 (1440564872257470464)]} 0 9
   [junit4]   2> 3982605 T8024 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[337]} 0 22
   [junit4]   2> 3982651 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[338 (1440564872314093568)]} 0 4
   [junit4]   2> 3982664 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[338 (1440564872323530752)]} 0 1
   [junit4]   2> 3982666 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[338 (1440564872323530752)]} 0 9
   [junit4]   2> 3982700 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[328 (-1440564872367570944)]} 0 1
   [junit4]   2> 3982711 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[328 (-1440564872375959552)]} 0 1
   [junit4]   2> 3982713 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[328 (-1440564872375959552)]} 0 6
   [junit4]   2> 3982715 T8027 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[328]} 0 11
   [junit4]   2> 3982719 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3982722 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[339 (1440564872390639616)]} 0 2
   [junit4]   2> 3982723 T8013 oasc.Overseer$ClusterStateUpdater.updateShardState Update shard state invoked for collection: collection1
   [junit4]   2> 3982724 T8013 oasc.Overseer$ClusterStateUpdater.updateShardState Update shard state shard1 to inactive
   [junit4]   2> 3982724 T8013 oasc.Overseer$ClusterStateUpdater.updateShardState Update shard state shard1_0 to active
   [junit4]   2> 3982725 T8013 oasc.Overseer$ClusterStateUpdater.updateShardState Update shard state shard1_1 to active
   [junit4]   2> 3982732 T8026 C1873 P51255 oasu.SolrCmdDistributor.syncAdd SYNCADD on StdNode: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ : 339
   [junit4]   2> 3982742 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[339 (1440564872401125376)]} 0 2
   [junit4]   2> 3982744 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&distrib.from.parent=shard1&version=2} {add=[339 (1440564872401125376)]} 0 9
   [junit4]   2> 3982745 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3982746 T8019 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3982746 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3982746 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3982746 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3982746 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3982751 T8056 C1872 P51261 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[339 (1440564872401125376)]} 0 1
   [junit4]   2> 3982752 T8026 C1873 P51255 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[339 (1440564872401125376)]} 0 21
   [junit4]   2> 3982753 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[339]} 0 25
   [junit4]   1> client port:0.0.0.0/0.0.0.0:0
   [junit4]   1> / (2)
   [junit4]   1> DATA:
   [junit4]   1>     
   [junit4]   1>  /solr (7)
   [junit4]   1>   /solr/configs (1)
   [junit4]   1>    /solr/configs/conf1 (9)
   [junit4]   1>     /solr/configs/conf1/mapping-ISOLatin1Accent.txt (0)
   [junit4]   1>     DATA: ...supressed...
   [junit4]   1>     /solr/configs/conf1/currency.xml (0)
   [junit4]   1>     DATA: ...supressed...
   [junit4]   1>     /solr/configs/conf1/protwords.txt (0)
   [junit4]   1>     DATA: ...supressed...
   [junit4]   1>     /solr/configs/conf1/synonyms.txt (0)
   [junit4]   1>     DATA: ...supressed...
   [junit4]   1>     /solr/configs/conf1/old_synonyms.txt (0)
   [junit4]   1>     DATA: ...supressed...
   [junit4]   1>     /solr/configs/conf1/solrconfig.xml (0)
   [junit4]   1>     DATA: ...supressed...
   [junit4]   1>     /solr/configs/conf1/stopwords.txt (0)
   [junit4]   1>     DATA: ...supressed...
   [junit4]   1>     /solr/configs/conf1/open-exchange-rates.json (0)
   [junit4]   1>     DATA:
   [junit4]   1>         {
   [junit4]   1>           "disclaimer": "This data is not real, it was synthetically created to match currency.xml.  It is modeled after the data format available from openexchangerates.org.  See https://openexchangerates.org/documentation for details",
   [junit4]   1>           "license": "http://www.apache.org/licenses/LICENSE-2.0",
   [junit4]   1>           "timestamp": 1332070464,
   [junit4]   1>         
   [junit4]   1>         
   [junit4]   1>           "IMPORTANT NOTE": "In order for tests to work, this data must be kept in sync with ./currency.xml",
   [junit4]   1>         
   [junit4]   1>         
   [junit4]   1>           "base": "USD",
   [junit4]   1>           "rates": {
   [junit4]   1>             "USD": 1,
   [junit4]   1>             "JPY": 81.29,
   [junit4]   1>             "EUR": 2.5,
   [junit4]   1>             "GBP": 0.5,
   [junit4]   1>             "MXN": 2.0
   [junit4]   1>           }
   [junit4]   1>         }
   [junit4]   1>         
   [junit4]   1>     /solr/configs/conf1/schema.xml (0)
   [junit4]   1>     DATA: ...supressed...
   [junit4]   1>   /solr/clusterstate.json (0)
   [junit4]   1>   DATA:
   [junit4]   1>       {
   [junit4]   1>         "collection1":{
   [junit4]   1>           "shards":{
   [junit4]   1>             "shard1":{
   [junit4]   1>               "range":"80000000-ffffffff",
   [junit4]   1>               "state":"active",
   [junit4]   1>               "replicas":{
   [junit4]   1>                 "core_node1":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1",
   [junit4]   1>                   "node_name":"127.0.0.1:51255_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51255/qb",
   [junit4]   1>                   "leader":"true"},
   [junit4]   1>                 "core_node3":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1",
   [junit4]   1>                   "node_name":"127.0.0.1:51261_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51261/qb"}}},
   [junit4]   1>             "shard2":{
   [junit4]   1>               "range":"0-7fffffff",
   [junit4]   1>               "state":"active",
   [junit4]   1>               "replicas":{
   [junit4]   1>                 "core_node2":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1",
   [junit4]   1>                   "node_name":"127.0.0.1:51258_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51258/qb",
   [junit4]   1>                   "leader":"true"},
   [junit4]   1>                 "core_node4":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1",
   [junit4]   1>                   "node_name":"127.0.0.1:51265_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51265/qb"}}},
   [junit4]   1>             "shard1_0":{
   [junit4]   1>               "range":"80000000-bfffffff",
   [junit4]   1>               "state":"construction",
   [junit4]   1>               "replicas":{
   [junit4]   1>                 "core_node5":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1_shard1_0_replica1",
   [junit4]   1>                   "node_name":"127.0.0.1:51255_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51255/qb",
   [junit4]   1>                   "leader":"true"},
   [junit4]   1>                 "core_node7":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1_shard1_0_replica2",
   [junit4]   1>                   "node_name":"127.0.0.1:51258_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51258/qb"}}},
   [junit4]   1>             "shard1_1":{
   [junit4]   1>               "range":"c0000000-ffffffff",
   [junit4]   1>               "state":"construction",
   [junit4]   1>               "replicas":{
   [junit4]   1>                 "core_node6":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1_shard1_1_replica1",
   [junit4]   1>                   "node_name":"127.0.0.1:51255_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51255/qb",
   [junit4]   1>                   "leader":"true"},
   [junit4]   1>                 "core_node8":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1_shard1_1_replica2",
   [junit4]   1>                   "node_name":"127.0.0.1:51251_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51251/qb"}}}},
   [junit4]   1>           "router":"compositeId"},
   [junit4]   1>         "control_collection":{
   [junit4]   1>           "shards":{"shard1":{
   [junit4]   1>               "range":"80000000-7fffffff",
   [junit4]   1>               "state":"active",
   [junit4]   1>               "replicas":{"core_node1":{
   [junit4]   1>                   "state":"active",
   [junit4]   1>                   "core":"collection1",
   [junit4]   1>                   "node_name":"127.0.0.1:51251_qb",
   [junit4]   1>                   "base_url":"http://127.0.0.1:51251/qb",
   [junit4]   1>                   "leader":"true"}}}},
   [junit4]   1>           "router":"compositeId"}}
   [junit4]   1>   /solr/aliases.json (0)
   [junit4]   1>   /solr/live_nodes (5)
   [junit4]   1>    /solr/live_nodes/127.0.0.1:51255_qb (0)
   [junit4]   1>    /solr/live_nodes/127.0.0.1:51261_qb (0)
   [junit4]   1>    /solr/live_nodes/127.0.0.1:51258_qb (0)
   [junit4]   1>    /solr/live_nodes/127.0.0.1:51265_qb (0)
   [junit4]   1>    /solr/live_nodes/127.0.0.1:51251_qb (0)
   [junit4]   1>   /solr/overseer (3)
   [junit4]   1>   DATA:
   [junit4]   1>       
   [junit4]   1>    /solr/overseer/queue (1)
   [junit4]   1>     /solr/overseer/queue/qn-0000000034 (0)
   [junit4]   1>     DATA:
   [junit4]   1>         {
   [junit4]   1>           "shard1":"inactive",
   [junit4]   1>           "operation":"updateshardstate",
   [junit4]   1>           "shard1_0":"active",
   [junit4]   1>           "collection":"collection1",
   [junit4]   1>           "shard1_1":"active"}
   [junit4]   1>    /solr/overseer/queue-work (0)
   [junit4]   1>    /solr/overseer/collection-queue-work (0)
   [junit4]   1>   /solr/collections (2)
   [junit4]   1>    /solr/collections/collection1 (3)
   [junit4]   1>    DATA:
   [junit4]   1>        {"configName":"conf1"}
   [junit4]   1>     /solr/collections/collection1/shards (0)
   [junit4]   1>     /solr/collections/collection1/leader_elect (4)
   [junit4]   1>      /solr/collections/collection1/leader_elect/shard1 (1)
   [junit4]   1>       /solr/collections/collection1/leader_elect/shard1/election (2)
   [junit4]   1>        /solr/collections/collection1/leader_elect/shard1/election/90035301003689990-core_node1-n_0000000000 (0)
   [junit4]   1>        /solr/collections/collection1/leader_elect/shard1/election/90035301003689994-core_node3-n_0000000001 (0)
   [junit4]   1>      /solr/collections/collection1/leader_elect/shard1_0 (1)
   [junit4]   1>       /solr/collections/collection1/leader_elect/shard1_0/election (2)
   [junit4]   1>        /solr/collections/collection1/leader_elect/shard1_0/election/90035301003689990-core_node5-n_0000000000 (0)
   [junit4]   1>        /solr/collections/collection1/leader_elect/shard1_0/election/90035301003689992-core_node7-n_0000000001 (0)
   [junit4]   1>      /solr/collections/collection1/leader_elect/shard1_1 (1)
   [junit4]   1>       /solr/collections/collection1/leader_elect/shard1_1/election (2)
   [junit4]   1>        /solr/collections/collection1/leader_elect/shard1_1/election/90035301003689987-core_node8-n_0000000001 (0)
   [junit4]   1>        /solr/collections/collection1/leader_elect/shard1_1/election/90035301003689990-core_node6-n_0000000000 (0)
   [junit4]   1>      /solr/collections/collection1/leader_elect/shard2 (1)
   [junit4]   1>       /solr/collections/collection1/leader_elect/shard2/election (2)
   [junit4]   1>        /solr/collections/collection1/leader_elect/shard2/election/90035301003689992-core_node2-n_0000000000 (0)
   [junit4]   1>        /solr/collections/collection1/leader_elect/shard2/election/90035301003689996-core_node4-n_0000000001 (0)
   [junit4]   1>     /solr/collections/collection1/leaders (4)
   [junit4]   1>      /solr/collections/collection1/leaders/shard1 (0)
   [junit4]   1>      DATA:
   [junit4]   1>          {
   [junit4]   1>            "core":"collection1",
   [junit4]   1>            "node_name":"127.0.0.1:51255_qb",
   [junit4]   1>            "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   1>      /solr/collections/collection1/leaders/shard1_0 (0)
   [junit4]   1>      DATA:
   [junit4]   1>          {
   [junit4]   1>            "core":"collection1_shard1_0_replica1",
   [junit4]   1>            "node_name":"127.0.0.1:51255_qb",
   [junit4]   1>            "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   1>      /solr/collections/collection1/leaders/shard1_1 (0)
   [junit4]   1>      DATA:
   [junit4]   1>          {
   [junit4]   1>            "core":"collection1_shard1_1_replica1",
   [junit4]   1>            "node_name":"127.0.0.1:51255_qb",
   [junit4]   1>            "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   1>      /solr/collections/collection1/leaders/shard2 (0)
   [junit4]   1>      DATA:
   [junit4]   1>          {
   [junit4]   1>            "core":"collection1",
   [junit4]   1>            "node_name":"127.0.0.1:51258_qb",
   [junit4]   1>            "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   1>    /solr/collections/control_collection (3)
   [junit4]   1>    DATA:
   [junit4]   1>        {"configName":"conf1"}
   [junit4]   1>     /solr/collections/control_collection/shards (0)
   [junit4]   1>     /solr/collections/control_collection/leader_elect (1)
   [junit4]   1>      /solr/collections/control_collection/leader_elect/shard1 (1)
   [junit4]   1>       /solr/collections/control_collection/leader_elect/shard1/election (1)
   [junit4]   1>        /solr/collections/control_collection/leader_elect/shard1/election/90035301003689987-core_node1-n_0000000000 (0)
   [junit4]   1>     /solr/collections/control_collection/leaders (1)
   [junit4]   1>      /solr/collections/control_collection/leaders/shard1 (0)
   [junit4]   1>      DATA:
   [junit4]   1>          {
   [junit4]   2> 3982792 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[340 (1440564872465088512)]} 0 0
   [junit4]   2> 3982805 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[340 (1440564872468234240)]} 0 3
   [junit4]   2> 3982806 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[340 (1440564872468234240)]} 0 11
   [junit4]   2> 3982841 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[341 (1440564872515420160)]} 0 1
   [junit4]   2> 3982851 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[341 (1440564872522760192)]} 0 1
   [junit4]   2> 3982852 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[341 (1440564872522760192)]} 0 5
   [junit4]   2> 3982853 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[341]} 0 9
   [junit4]   2> 3982886 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[317 (-1440564872563654656)]} 0 0
   [junit4]   2> 3982896 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[317 (-1440564872569946112)]} 0 0
   [junit4]   2> 3982898 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[317 (-1440564872569946112)]} 0 6
   [junit4]   2> 3982899 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[317]} 0 10
   [junit4]   2> 3982903 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[342 (1440564872580431872)]} 0 1
   [junit4]   2> 3982916 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[342 (1440564872591966208)]} 0 0
   [junit4]   2> 3982918 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[342 (1440564872591966208)]} 0 5
   [junit4]   2> 3982919 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[342]} 0 9
   [junit4]   2> 3982954 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[206 (-1440564872634957824)]} 0 1
   [junit4]   2> 3982963 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[206 (-1440564872641249280)]} 0 1
   [junit4]   2> 3982967 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[206 (-1440564872641249280)]} 0 8
   [junit4]   2> 3982973 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[343 (1440564872653832192)]} 0 1
   [junit4]   2> 3982983 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[343 (1440564872661172224)]} 0 1
   [junit4]   2> 3982984 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[343 (1440564872661172224)]} 0 5
   [junit4]   2> 3982985 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[343]} 0 9
   [junit4]   2> 3983018 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[332 (-1440564872702066688)]} 0 1
   [junit4]   2> 3983028 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[332 (-1440564872710455296)]} 0 0
   [junit4]   2> 3983030 T8024 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[332 (-1440564872710455296)]} 0 4
   [junit4]   2> 3983031 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[332]} 0 10
   [junit4]   2> 3983036 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[344 (1440564872719892480)]} 0 1
   [junit4]   2> 3983051 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[344 (1440564872727232512)]} 0 1
   [junit4]   2> 3983057 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[344 (1440564872727232512)]} 0 16
   [junit4]   2> 3983061 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[344]} 0 22
   [junit4]   2> 3983094 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[339 (-1440564872781758464)]} 0 0
   [junit4]   2> 3983103 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[339 (-1440564872788049920)]} 0 1
   [junit4]   2> 3983105 T8026 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[339 (-1440564872788049920)]} 0 5
   [junit4]   2> 3983111 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[339]} 0 13
   [junit4]   2> 3983115 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[345 (1440564872803778560)]} 0 0
   [junit4]   2> 3983125 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[345 (1440564872810070016)]} 0 1
   [junit4]   2> 3983127 T8024 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[345 (1440564872810070016)]} 0 6
   [junit4]   2> 3983127 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[345]} 0 9
   [junit4]   2> 3983160 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[346 (1440564872850964480)]} 0 1
   [junit4]   2> 3983171 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[346 (1440564872859353088)]} 0 0
   [junit4]   2> 3983173 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[346 (1440564872859353088)]} 0 6
   [junit4]   2> 3983174 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[346]} 0 9
   [junit4]   2> 3983214 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[347 (1440564872906539008)]} 0 1
   [junit4]   2> 3983225 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[347 (1440564872914927616)]} 0 1
   [junit4]   2> 3983226 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[347 (1440564872914927616)]} 0 5
   [junit4]   2> 3983227 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[347]} 0 10
   [junit4]   2> 3983260 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[235 (-1440564872955822080)]} 0 1
   [junit4]   2> 3983269 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[235 (-1440564872962113536)]} 0 1
   [junit4]   2> 3983271 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[235 (-1440564872962113536)]} 0 5
   [junit4]   2> 3983272 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[235]} 0 9
   [junit4]   2> 3983277 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[348 (1440564872972599296)]} 0 1
   [junit4]   2> 3983284 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[348 (1440564872976793600)]} 0 1
   [junit4]   2> 3983285 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[348 (1440564872976793600)]} 0 5
   [junit4]   2> 3983320 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[349 (1440564873018736640)]} 0 1
   [junit4]   2> 3983327 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[349 (1440564873021882368)]} 0 1
   [junit4]   2> 3983328 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[349 (1440564873021882368)]} 0 5
   [junit4]   2> 3983362 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[350 (1440564873061728256)]} 0 0
   [junit4]   2> 3983369 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[350 (1440564873065922560)]} 0 0
   [junit4]   2> 3983372 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[350 (1440564873065922560)]} 0 7
   [junit4]   2> 3983406 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[312 (-1440564873107865600)]} 0 1
   [junit4]   2> 3983415 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[312 (-1440564873115205632)]} 0 1
   [junit4]   2> 3983416 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[312 (-1440564873115205632)]} 0 4
   [junit4]   2> 3983419 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[312]} 0 10
   [junit4]   2> 3983427 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[351 (1440564873129885696)]} 0 0
   [junit4]   2> 3983436 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[351 (1440564873137225728)]} 0 1
   [junit4]   2> 3983437 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[351 (1440564873137225728)]} 0 5
   [junit4]   2> 3983438 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[351]} 0 9
   [junit4]   2> 3983471 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[258 (-1440564873176023040)]} 0 1
   [junit4]   2> 3983478 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[258 (-1440564873181265920)]} 0 0
   [junit4]   2> 3983482 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[258 (-1440564873181265920)]} 0 7
   [junit4]   2> 3983492 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[352 (1440564873198043136)]} 0 1
   [junit4]   2> 3983503 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[352 (1440564873206431744)]} 0 1
   [junit4]   2> 3983505 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[352 (1440564873206431744)]} 0 6
   [junit4]   2> 3983506 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[352]} 0 9
   [junit4]   2> 3983540 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[263 (-1440564873249423360)]} 0 1
   [junit4]   2> 3983549 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[263 (-1440564873256763392)]} 0 0
   [junit4]   2> 3983551 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[263 (-1440564873256763392)]} 0 5
   [junit4]   2> 3983553 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[263]} 0 10
   [junit4]   2> 3983561 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[353 (1440564873269346304)]} 0 3
   [junit4]   2> 3983571 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[353 (1440564873277734912)]} 0 1
   [junit4]   2> 3983572 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[353 (1440564873277734912)]} 0 5
   [junit4]   2> 3983606 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[354 (1440564873317580800)]} 0 1
   [junit4]   2> 3983624 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[354 (1440564873332260864)]} 0 1
   [junit4]   2> 3983625 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[354 (1440564873332260864)]} 0 7
   [junit4]   2> 3983626 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[354]} 0 14
   [junit4]   2> 3983659 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[298 (-1440564873373155328)]} 0 1
   [junit4]   2> 3983665 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[298 (-1440564873377349632)]} 0 0
   [junit4]   2> 3983667 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[298 (-1440564873377349632)]} 0 5
   [junit4]   2> 3983671 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[355 (1440564873386786816)]} 0 0
   [junit4]   2> 3983687 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[355 (1440564873393078272)]} 0 5
   [junit4]   2> 3983689 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[355 (1440564873393078272)]} 0 12
   [junit4]   2> 3983690 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[355]} 0 15
   [junit4]   2> 3983724 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[356 (1440564873441312768)]} 0 1
   [junit4]   2> 3983731 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[356 (1440564873445507072)]} 0 1
   [junit4]   2> 3983732 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[356 (1440564873445507072)]} 0 5
   [junit4]   2> 3983765 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[357 (1440564873485352960)]} 0 0
   [junit4]   2> 3983775 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[357 (1440564873492692992)]} 0 1
   [junit4]   2> 3983776 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[357 (1440564873492692992)]} 0 5
   [junit4]   2> 3983777 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[357]} 0 8
   [junit4]   2> 3983813 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[323 (-1440564873535684608)]} 0 0
   [junit4]   2> 3983821 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[323 (-1440564873539878912)]} 0 1
   [junit4]   2> 3983822 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[323 (-1440564873539878912)]} 0 5
   [junit4]   2> 3983826 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[358 (1440564873549316096)]} 0 0
   [junit4]   2> 3983833 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[358 (1440564873553510400)]} 0 0
   [junit4]   2> 3983835 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[358 (1440564873553510400)]} 0 5
   [junit4]   2> 3983872 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[359 (1440564873593356288)]} 0 5
   [junit4]   2> 3983882 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[359 (1440564873604890624)]} 0 0
   [junit4]   2> 3983883 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[359 (1440564873604890624)]} 0 4
   [junit4]   2> 3983884 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[359]} 0 7
   [junit4]   2> 3983917 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[348 (-1440564873644736512)]} 0 0
   [junit4]   2> 3983923 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[348 (-1440564873648930816)]} 0 0
   [junit4]   2> 3983925 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[348 (-1440564873648930816)]} 0 5
   [junit4]   2> 3983931 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[360 (1440564873657319424)]} 0 1
   [junit4]   2> 3983942 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[360 (1440564873665708032)]} 0 3
   [junit4]   2> 3983947 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[360 (1440564873665708032)]} 0 10
   [junit4]   2> 3983948 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[360]} 0 14
   [junit4]   2> 3983981 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[342 (-1440564873711845376)]} 0 0
   [junit4]   2> 3983990 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[342 (-1440564873718136832)]} 0 0
   [junit4]   2> 3983991 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[342 (-1440564873718136832)]} 0 4
   [junit4]   2> 3983992 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[342]} 0 8
   [junit4]   2> 3983996 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[361 (1440564873726525440)]} 0 1
   [junit4]   2> 3984003 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[361 (1440564873731768320)]} 0 0
   [junit4]   2> 3984005 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[361 (1440564873731768320)]} 0 5
   [junit4]   2> 3984038 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[288 (-1440564873771614208)]} 0 0
   [junit4]   2> 3984045 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[288 (-1440564873775808512)]} 0 0
   [junit4]   2> 3984047 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[288 (-1440564873775808512)]} 0 5
   [junit4]   2> 3984050 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[362 (1440564873784197120)]} 0 0
   [junit4]   2> 3984060 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[362 (1440564873789440000)]} 0 1
   [junit4]   2> 3984061 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[362 (1440564873789440000)]} 0 6
   [junit4]   2> 3984062 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[362]} 0 9
   [junit4]   2> 3984094 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[203 (-1440564873830334464)]} 0 0
   [junit4]   2> 3984105 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[203 (-1440564873838723072)]} 0 0
   [junit4]   2> 3984107 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[203 (-1440564873838723072)]} 0 5
   [junit4]   2> 3984109 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[203]} 0 9
   [junit4]   2> 3984113 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[363 (1440564873850257408)]} 0 1
   [junit4]   2> 3984124 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[363 (1440564873857597440)]} 0 1
   [junit4]   2> 3984126 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[363 (1440564873857597440)]} 0 6
   [junit4]   2> 3984127 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[363]} 0 10
   [junit4]   2> 3984162 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[364 (1440564873900589056)]} 0 1
   [junit4]   2> 3984174 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[364 (1440564873908977664)]} 0 1
   [junit4]   2> 3984175 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[364 (1440564873908977664)]} 0 6
   [junit4]   2> 3984177 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[364]} 0 11
   [junit4]   2> 3984216 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[282 (-1440564873958260736)]} 0 1
   [junit4]   2> 3984229 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[282 (-1440564873968746496)]} 0 0
   [junit4]   2> 3984231 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[282 (-1440564873968746496)]} 0 5
   [junit4]   2> 3984233 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[282]} 0 10
   [junit4]   2> 3984242 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[365 (1440564873984475136)]} 0 1
   [junit4]   2> 3984253 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[365 (1440564873992863744)]} 0 1
   [junit4]   2> 3984255 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[365 (1440564873992863744)]} 0 6
   [junit4]   2> 3984256 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[365]} 0 10
   [junit4]   2> 3984292 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[366 (1440564874035855360)]} 0 3
   [junit4]   2> 3984308 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[366 (1440564874048438272)]} 0 2
   [junit4]   2> 3984311 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[366 (1440564874048438272)]} 0 9
   [junit4]   2> 3984318 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[366]} 0 19
   [junit4]   2> 3984355 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[280 (-1440564874104012800)]} 0 1
   [junit4]   2> 3984364 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[280 (-1440564874110304256)]} 0 1
   [junit4]   2> 3984366 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[280 (-1440564874110304256)]} 0 5
   [junit4]   2> 3984367 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[280]} 0 9
   [junit4]   2> 3984374 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[367 (1440564874123935744)]} 0 1
   [junit4]   2> 3984381 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[367 (1440564874128130048)]} 0 1
   [junit4]   2> 3984383 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[367 (1440564874128130048)]} 0 5
   [junit4]   2> 3984419 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[163 (-1440564874171121664)]} 0 1
   [junit4]   2> 3984431 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[163 (-1440564874180558848)]} 0 1
   [junit4]   2> 3984432 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[163 (-1440564874180558848)]} 0 5
   [junit4]   2> 3984434 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[163]} 0 11
   [junit4]   2> 3984439 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[368 (1440564874191044608)]} 0 1
   [junit4]   2> 3984446 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[368 (1440564874195238912)]} 0 1
   [junit4]   2> 3984448 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[368 (1440564874195238912)]} 0 6
   [junit4]   2> 3984481 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[156 (-1440564874235084800)]} 0 1
   [junit4]   2> 3984488 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[156 (-1440564874240327680)]} 0 1
   [junit4]   2> 3984489 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[156 (-1440564874240327680)]} 0 4
   [junit4]   2> 3984493 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[369 (1440564874248716288)]} 0 1
   [junit4]   2> 3984503 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[369 (1440564874255007744)]} 0 0
   [junit4]   2> 3984504 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[369 (1440564874255007744)]} 0 5
   [junit4]   2> 3984505 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[369]} 0 9
   [junit4]   2> 3984539 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[357 (-1440564874295902208)]} 0 0
   [junit4]   2> 3984551 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[357 (-1440564874306387968)]} 0 0
   [junit4]   2> 3984553 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[357 (-1440564874306387968)]} 0 5
   [junit4]   2> 3984554 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[357]} 0 9
   [junit4]   2> 3984559 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[370 (1440564874316873728)]} 0 1
   [junit4]   2> 3984569 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[370 (1440564874324213760)]} 0 1
   [junit4]   2> 3984570 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[370 (1440564874324213760)]} 0 5
   [junit4]   2> 3984572 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[370]} 0 10
   [junit4]   2> 3984611 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[176 (-1440564874372448256)]} 0 1
   [junit4]   2> 3984622 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[176 (-1440564874379788288)]} 0 1
   [junit4]   2> 3984623 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[176 (-1440564874379788288)]} 0 5
   [junit4]   2> 3984624 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[176]} 0 9
   [junit4]   2> 3984628 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[371 (1440564874389225472)]} 0 1
   [junit4]   2> 3984641 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[371 (1440564874397614080)]} 0 2
   [junit4]   2> 3984646 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[371 (1440564874397614080)]} 0 12
   [junit4]   2> 3984650 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[371]} 0 19
   [junit4]   2> 3984684 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[372 (1440564874447945728)]} 0 1
   [junit4]   2> 3984694 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[372 (1440564874455285760)]} 0 1
   [junit4]   2> 3984695 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[372 (1440564874455285760)]} 0 5
   [junit4]   2> 3984697 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[372]} 0 9
   [junit4]   2> 3984730 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[216 (-1440564874497228800)]} 0 1
   [junit4]   2> 3984740 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[216 (-1440564874503520256)]} 0 1
   [junit4]   2> 3984741 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[216 (-1440564874503520256)]} 0 5
   [junit4]   2> 3984742 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[216]} 0 8
   [junit4]   2> 3984746 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[373 (1440564874514006016)]} 0 0
   [junit4]   2> 3984754 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[373 (1440564874518200320)]} 0 0
   [junit4]   2> 3984756 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[373 (1440564874518200320)]} 0 6
   [junit4]   2> 3984792 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[257 (-1440564874562240512)]} 0 0
   [junit4]   2> 3984808 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[257 (-1440564874574823424)]} 0 1
   [junit4]   2> 3984809 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[257 (-1440564874574823424)]} 0 5
   [junit4]   2> 3984810 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[257]} 0 9
   [junit4]   2> 3984815 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[374 (1440564874585309184)]} 0 1
   [junit4]   2> 3984822 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[374 (1440564874589503488)]} 0 1
   [junit4]   2> 3984823 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[374 (1440564874589503488)]} 0 5
   [junit4]   2> 3984856 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[375 (1440564874629349376)]} 0 1
   [junit4]   2> 3984869 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[375 (1440564874638786560)]} 0 1
   [junit4]   2> 3984871 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[375 (1440564874638786560)]} 0 6
   [junit4]   2> 3984872 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[375]} 0 9
   [junit4]   2> 3984905 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[314 (-1440564874680729600)]} 0 1
   [junit4]   2> 3984922 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[314 (-1440564874689118208)]} 0 0
   [junit4]   2> 3984924 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[314 (-1440564874689118208)]} 0 12
   [junit4]   2> 3984925 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[314]} 0 16
   [junit4]   2> 3984930 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[376 (1440564874705895424)]} 0 1
   [junit4]   2> 3984939 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[376 (1440564874713235456)]} 0 0
   [junit4]   2> 3984941 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[376 (1440564874713235456)]} 0 5
   [junit4]   2> 3984942 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[376]} 0 8
   [junit4]   2> 3984975 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[220 (-1440564874754129920)]} 0 0
   [junit4]   2> 3985000 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[220 (-1440564874768809984)]} 0 0
   [junit4]   2> 3985005 T8024 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[220 (-1440564874768809984)]} 0 17
   [junit4]   2> 3985007 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[220]} 0 26
   [junit4]   2> 3985020 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[377 (1440564874800267264)]} 0 1
   [junit4]   2> 3985033 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[377 (1440564874810753024)]} 0 1
   [junit4]   2> 3985034 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[377 (1440564874810753024)]} 0 6
   [junit4]   2> 3985071 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[319 (-1440564874853744640)]} 0 1
   [junit4]   2> 3985085 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[319 (-1440564874866327552)]} 0 0
   [junit4]   2> 3985086 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[319 (-1440564874866327552)]} 0 4
   [junit4]   2> 3985087 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[319]} 0 8
   [junit4]   2> 3985099 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[378 (1440564874884153344)]} 0 1
   [junit4]   2> 3985135 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[378 (1440564874911416320)]} 0 5
   [junit4]   2> 3985145 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[378 (1440564874911416320)]} 0 21
   [junit4]   2> 3985147 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[378]} 0 33
   [junit4]   2> 3985187 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[379 (1440564874976428032)]} 0 1
   [junit4]   2> 3985203 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[379 (1440564874990059520)]} 0 1
   [junit4]   2> 3985214 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[379 (1440564874990059520)]} 0 15
   [junit4]   2> 3985226 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[379]} 0 33
   [junit4]   2> 3985265 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[309 (-1440564875057168384)]} 0 1
   [junit4]   2> 3985283 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[309 (-1440564875061362688)]} 0 3
   [junit4]   2> 3985285 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[309 (-1440564875061362688)]} 0 17
   [junit4]   2> 3985303 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[380 (1440564875097014272)]} 0 1
   [junit4]   2> 3985329 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[380 (1440564875117985792)]} 0 1
   [junit4]   2> 3985330 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[380 (1440564875117985792)]} 0 12
   [junit4]   2> 3985368 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[381 (1440564875166220288)]} 0 1
   [junit4]   2> 3985396 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[381 (1440564875187191808)]} 0 2
   [junit4]   2> 3985397 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[381 (1440564875187191808)]} 0 10
   [junit4]   2> 3985399 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[381]} 0 16
   [junit4]   2> 3985433 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[382 (1440564875233329152)]} 0 1
   [junit4]   2> 3985443 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[382 (1440564875241717760)]} 0 1
   [junit4]   2> 3985444 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[382 (1440564875241717760)]} 0 4
   [junit4]   2> 3985446 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[382]} 0 9
   [junit4]   2> 3985479 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[305 (-1440564875282612224)]} 0 0
   [junit4]   2> 3985489 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[305 (-1440564875291000832)]} 0 0
   [junit4]   2> 3985491 T8024 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[305 (-1440564875291000832)]} 0 5
   [junit4]   2> 3985493 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[305]} 0 9
   [junit4]   2> 3985498 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[383 (1440564875301486592)]} 0 1
   [junit4]   2> 3985508 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[383 (1440564875308826624)]} 0 1
   [junit4]   2> 3985510 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[383 (1440564875308826624)]} 0 6
   [junit4]   2> 3985512 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[383]} 0 10
   [junit4]   2> 3985547 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[384 (1440564875353915392)]} 0 1
   [junit4]   2> 3985558 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[384 (1440564875361255424)]} 0 1
   [junit4]   2> 3985559 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[384 (1440564875361255424)]} 0 5
   [junit4]   2> 3985560 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[384]} 0 9
   [junit4]   2> 3985597 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[177 (-1440564875406344192)]} 0 0
   [junit4]   2> 3985607 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[177 (-1440564875413684224)]} 0 0
   [junit4]   2> 3985609 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[177 (-1440564875413684224)]} 0 5
   [junit4]   2> 3985610 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[177]} 0 9
   [junit4]   2> 3985615 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[385 (1440564875424169984)]} 0 1
   [junit4]   2> 3985623 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[385 (1440564875429412864)]} 0 1
   [junit4]   2> 3985624 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[385 (1440564875429412864)]} 0 5
   [junit4]   2> 3985661 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[175 (-1440564875472404480)]} 0 2
   [junit4]   2> 3985671 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[175 (-1440564875479744512)]} 0 1
   [junit4]   2> 3985672 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[175 (-1440564875479744512)]} 0 6
   [junit4]   2> 3985677 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[386 (1440564875489181696)]} 0 1
   [junit4]   2> 3985688 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[386 (1440564875498618880)]} 0 1
   [junit4]   2> 3985690 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[386 (1440564875498618880)]} 0 5
   [junit4]   2> 3985725 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[387 (1440564875538464768)]} 0 3
   [junit4]   2> 3985741 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[387 (1440564875552096256)]} 0 0
   [junit4]   2> 3985742 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[387 (1440564875552096256)]} 0 6
   [junit4]   2> 3985776 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[388 (1440564875592990720)]} 0 1
   [junit4]   2> 3985782 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[388 (1440564875597185024)]} 0 1
   [junit4]   2> 3985783 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[388 (1440564875597185024)]} 0 4
   [junit4]   2> 3985815 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[240 (-1440564875634933760)]} 0 0
   [junit4]   2> 3985821 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[240 (-1440564875638079488)]} 0 0
   [junit4]   2> 3985822 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[240 (-1440564875638079488)]} 0 4
   [junit4]   2> 3985826 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[389 (1440564875646468096)]} 0 1
   [junit4]   2> 3985836 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[389 (1440564875652759552)]} 0 1
   [junit4]   2> 3985837 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[389 (1440564875652759552)]} 0 5
   [junit4]   2> 3985839 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[389]} 0 10
   [junit4]   2> 3985872 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[346 (-1440564875694702592)]} 0 0
   [junit4]   2> 3985881 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[346 (-1440564875700994048)]} 0 0
   [junit4]   2> 3985883 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[346 (-1440564875700994048)]} 0 5
   [junit4]   2> 3985884 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[346]} 0 8
   [junit4]   2> 3985888 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[390 (1440564875710431232)]} 0 1
   [junit4]   2> 3985898 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[390 (1440564875717771264)]} 0 1
   [junit4]   2> 3985899 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[390 (1440564875717771264)]} 0 5
   [junit4]   2> 3985901 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[390]} 0 10
   [junit4]   2> 3985934 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[391 (1440564875759714304)]} 0 1
   [junit4]   2> 3985946 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[391 (1440564875767054336)]} 0 0
   [junit4]   2> 3985948 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[391 (1440564875767054336)]} 0 8
   [junit4]   2> 3985948 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[391]} 0 10
   [junit4]   2> 3985981 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[392 (1440564875808997376)]} 0 0
   [junit4]   2> 3985994 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[392 (1440564875815288832)]} 0 1
   [junit4]   2> 3985995 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[392 (1440564875815288832)]} 0 8
   [junit4]   2> 3986032 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[393 (1440564875861426176)]} 0 1
   [junit4]   2> 3986039 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[393 (1440564875864571904)]} 0 1
   [junit4]   2> 3986041 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[393 (1440564875864571904)]} 0 7
   [junit4]   2> 3986078 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[151 (-1440564875909660672)]} 0 1
   [junit4]   2> 3986086 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[151 (-1440564875917000704)]} 0 0
   [junit4]   2> 3986088 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[151 (-1440564875917000704)]} 0 4
   [junit4]   2> 3986089 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[151]} 0 8
   [junit4]   2> 3986094 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[394 (1440564875926437888)]} 0 1
   [junit4]   2> 3986101 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[394 (1440564875931680768)]} 0 1
   [junit4]   2> 3986102 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[394 (1440564875931680768)]} 0 5
   [junit4]   2> 3986138 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[395 (1440564875972575232)]} 0 2
   [junit4]   2> 3986147 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[395 (1440564875979915264)]} 0 0
   [junit4]   2> 3986149 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[395 (1440564875979915264)]} 0 5
   [junit4]   2> 3986150 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[395]} 0 9
   [junit4]   2> 3986183 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[334 (-1440564876020809728)]} 0 0
   [junit4]   2> 3986193 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[334 (-1440564876028149760)]} 0 0
   [junit4]   2> 3986194 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[334 (-1440564876028149760)]} 0 5
   [junit4]   2> 3986197 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[334]} 0 10
   [junit4]   2> 3986203 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[396 (1440564876040732672)]} 0 1
   [junit4]   2> 3986217 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[396 (1440564876051218432)]} 0 1
   [junit4]   2> 3986218 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[396 (1440564876051218432)]} 0 6
   [junit4]   2> 3986220 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[396]} 0 11
   [junit4]   2> 3986257 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[356 (-1440564876098404352)]} 0 1
   [junit4]   2> 3986268 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[356 (-1440564876105744384)]} 0 1
   [junit4]   2> 3986269 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[356 (-1440564876105744384)]} 0 5
   [junit4]   2> 3986273 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[397 (1440564876114132992)]} 0 1
   [junit4]   2> 3986279 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[397 (1440564876118327296)]} 0 1
   [junit4]   2> 3986280 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[397 (1440564876118327296)]} 0 4
   [junit4]   2> 3986318 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[186 (-1440564876162367488)]} 0 0
   [junit4]   2> 3986331 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[186 (-1440564876172853248)]} 0 1
   [junit4]   2> 3986332 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[186 (-1440564876172853248)]} 0 4
   [junit4]   2> 3986333 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[186]} 0 11
   [junit4]   2> 3986338 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[398 (1440564876183339008)]} 0 0
   [junit4]   2> 3986345 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[398 (1440564876187533312)]} 0 0
   [junit4]   2> 3986347 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[398 (1440564876187533312)]} 0 5
   [junit4]   2> 3986384 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[399 (1440564876230524928)]} 0 1
   [junit4]   2> 3986393 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[399 (1440564876236816384)]} 0 1
   [junit4]   2> 3986394 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[399 (1440564876236816384)]} 0 5
   [junit4]   2> 3986428 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[400 (1440564876276662272)]} 0 1
   [junit4]   2> 3986437 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[400 (1440564876284002304)]} 0 0
   [junit4]   2> 3986439 T8026 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[400 (1440564876284002304)]} 0 5
   [junit4]   2> 3986440 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[400]} 0 9
   [junit4]   2> 3986473 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[213 (-1440564876324896768)]} 0 0
   [junit4]   2> 3986480 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[213 (-1440564876329091072)]} 0 0
   [junit4]   2> 3986484 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[213 (-1440564876329091072)]} 0 7
   [junit4]   2> 3986491 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[401 (1440564876342722560)]} 0 1
   [junit4]   2> 3986502 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[401 (1440564876352159744)]} 0 0
   [junit4]   2> 3986504 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[401 (1440564876352159744)]} 0 5
   [junit4]   2> 3986505 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[401]} 0 9
   [junit4]   2> 3986539 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[236 (-1440564876394102784)]} 0 0
   [junit4]   2> 3986550 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[236 (-1440564876403539968)]} 0 0
   [junit4]   2> 3986552 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[236 (-1440564876403539968)]} 0 5
   [junit4]   2> 3986553 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[236]} 0 8
   [junit4]   2> 3986557 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[402 (1440564876412977152)]} 0 1
   [junit4]   2> 3986564 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[402 (1440564876417171456)]} 0 1
   [junit4]   2> 3986565 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[402 (1440564876417171456)]} 0 4
   [junit4]   2> 3986598 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[318 (-1440564876455968768)]} 0 0
   [junit4]   2> 3986607 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[318 (-1440564876462260224)]} 0 0
   [junit4]   2> 3986610 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[318 (-1440564876462260224)]} 0 6
   [junit4]   2> 3986614 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[318]} 0 12
   [junit4]   2> 3986620 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[403 (1440564876477988864)]} 0 1
   [junit4]   2> 3986626 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[403 (1440564876482183168)]} 0 1
   [junit4]   2> 3986627 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[403 (1440564876482183168)]} 0 4
   [junit4]   2> 3986661 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[404 (1440564876520980480)]} 0 1
   [junit4]   2> 3986672 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[404 (1440564876529369088)]} 0 1
   [junit4]   2> 3986673 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[404 (1440564876529369088)]} 0 5
   [junit4]   2> 3986674 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[404]} 0 8
   [junit4]   2> 3986711 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[372 (-1440564876574457856)]} 0 1
   [junit4]   2> 3986726 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[372 (-1440564876588089344)]} 0 0
   [junit4]   2> 3986728 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[372 (-1440564876588089344)]} 0 4
   [junit4]   2> 3986729 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[372]} 0 8
   [junit4]   2> 3986733 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[405 (1440564876597526528)]} 0 0
   [junit4]   2> 3986742 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[405 (1440564876603817984)]} 0 0
   [junit4]   2> 3986744 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[405 (1440564876603817984)]} 0 5
   [junit4]   2> 3986746 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[405]} 0 10
   [junit4]   2> 3986783 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[406 (1440564876648906752)]} 0 2
   [junit4]   2> 3986800 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[406 (1440564876660441088)]} 0 1
   [junit4]   2> 3986801 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[406 (1440564876660441088)]} 0 8
   [junit4]   2> 3986833 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[407 (1440564876702384128)]} 0 0
   [junit4]   2> 3986848 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[407 (1440564876712869888)]} 0 0
   [junit4]   2> 3986849 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[407 (1440564876712869888)]} 0 8
   [junit4]   2> 3986850 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[407]} 0 13
   [junit4]   2> 3986884 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[274 (-1440564876755861504)]} 0 1
   [junit4]   2> 3986895 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[274 (-1440564876763201536)]} 0 1
   [junit4]   2> 3986897 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[274 (-1440564876763201536)]} 0 6
   [junit4]   2> 3986898 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[274]} 0 10
   [junit4]   2> 3986901 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[408 (1440564876773687296)]} 0 1
   [junit4]   2> 3986910 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[408 (1440564876779978752)]} 0 1
   [junit4]   2> 3986911 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[408 (1440564876779978752)]} 0 4
   [junit4]   2> 3986912 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[408]} 0 8
   [junit4]   2> 3986945 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[409 (1440564876819824640)]} 0 1
   [junit4]   2> 3986952 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[409 (1440564876824018944)]} 0 1
   [junit4]   2> 3986953 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[409 (1440564876824018944)]} 0 4
   [junit4]   2> 3986985 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[385 (-1440564876861767680)]} 0 0
   [junit4]   2> 3986994 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[385 (-1440564876865961984)]} 0 1
   [junit4]   2> 3986995 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[385 (-1440564876865961984)]} 0 6
   [junit4]   2> 3986999 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[410 (1440564876875399168)]} 0 1
   [junit4]   2> 3987009 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[410 (1440564876882739200)]} 0 1
   [junit4]   2> 3987011 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[410 (1440564876882739200)]} 0 6
   [junit4]   2> 3987013 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[410]} 0 11
   [junit4]   2> 3987048 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[289 (-1440564876927827968)]} 0 0
   [junit4]   2> 3987060 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[289 (-1440564876938313728)]} 0 0
   [junit4]   2> 3987062 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[289 (-1440564876938313728)]} 0 5
   [junit4]   2> 3987062 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[289]} 0 7
   [junit4]   2> 3987066 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[411 (1440564876946702336)]} 0 1
   [junit4]   2> 3987076 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[411 (1440564876952993792)]} 0 1
   [junit4]   2> 3987078 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[411 (1440564876952993792)]} 0 6
   [junit4]   2> 3987080 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[411]} 0 11
   [junit4]   2> 3987118 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[412 (1440564877000179712)]} 0 1
   [junit4]   2> 3987131 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[412 (1440564877010665472)]} 0 1
   [junit4]   2> 3987132 T8024 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[412 (1440564877010665472)]} 0 5
   [junit4]   2> 3987133 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[412]} 0 12
   [junit4]   2> 3987167 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[413 (1440564877052608512)]} 0 1
   [junit4]   2> 3987184 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[413 (1440564877064142848)]} 0 1
   [junit4]   2> 3987186 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[413 (1440564877064142848)]} 0 9
   [junit4]   2> 3987187 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[413]} 0 14
   [junit4]   2> 3987220 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[410 (-1440564877108183040)]} 0 0
   [junit4]   2> 3987229 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[410 (-1440564877114474496)]} 0 1
   [junit4]   2> 3987230 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[410 (-1440564877114474496)]} 0 4
   [junit4]   2> 3987232 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[410]} 0 8
   [junit4]   2> 3987235 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[414 (1440564877123911680)]} 0 0
   [junit4]   2> 3987249 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[414 (1440564877130203136)]} 0 1
   [junit4]   2> 3987250 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[414 (1440564877130203136)]} 0 9
   [junit4]   2> 3987252 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[414]} 0 14
   [junit4]   2> 3987285 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[414 (-1440564877176340480)]} 0 0
   [junit4]   2> 3987294 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[414 (-1440564877183680512)]} 0 0
   [junit4]   2> 3987296 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[414 (-1440564877183680512)]} 0 5
   [junit4]   2> 3987298 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[414]} 0 9
   [junit4]   2> 3987320 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[415 (1440564877211992064)]} 0 0
   [junit4]   2> 3987374 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[415 (1440564877228769280)]} 0 1
   [junit4]   2> 3987379 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[415 (1440564877228769280)]} 0 44
   [junit4]   2> 3987381 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[415]} 0 51
   [junit4]   2> 3987423 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[416 (1440564877319995392)]} 0 1
   [junit4]   2> 3987430 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[416 (1440564877325238272)]} 0 0
   [junit4]   2> 3987431 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[416 (1440564877325238272)]} 0 5
   [junit4]   2> 3987464 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[373 (-1440564877364035584)]} 0 0
   [junit4]   2> 3987477 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[373 (-1440564877374521344)]} 0 1
   [junit4]   2> 3987480 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[373 (-1440564877374521344)]} 0 7
   [junit4]   2> 3987489 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[417 (1440564877389201408)]} 0 2
   [junit4]   2> 3987502 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[417 (1440564877400735744)]} 0 1
   [junit4]   2> 3987503 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[417 (1440564877400735744)]} 0 4
   [junit4]   2> 3987504 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[417]} 0 8
   [junit4]   2> 3987545 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[316 (-1440564877448970240)]} 0 1
   [junit4]   2> 3987554 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[316 (-1440564877454213120)]} 0 1
   [junit4]   2> 3987556 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[316 (-1440564877454213120)]} 0 6
   [junit4]   2> 3987560 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[418 (1440564877464698880)]} 0 0
   [junit4]   2> 3987569 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[418 (1440564877470990336)]} 0 0
   [junit4]   2> 3987570 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[418 (1440564877470990336)]} 0 4
   [junit4]   2> 3987572 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[418]} 0 9
   [junit4]   2> 3987607 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[286 (-1440564877513981952)]} 0 1
   [junit4]   2> 3987617 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[286 (-1440564877520273408)]} 0 1
   [junit4]   2> 3987619 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[286 (-1440564877520273408)]} 0 6
   [junit4]   2> 3987620 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[286]} 0 10
   [junit4]   2> 3987624 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[419 (1440564877531807744)]} 0 1
   [junit4]   2> 3987633 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[419 (1440564877538099200)]} 0 0
   [junit4]   2> 3987635 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[419 (1440564877538099200)]} 0 5
   [junit4]   2> 3987636 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[419]} 0 9
   [junit4]   2> 3987668 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[391 (-1440564877577945088)]} 0 0
   [junit4]   2> 3987677 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[391 (-1440564877584236544)]} 0 0
   [junit4]   2> 3987679 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[391 (-1440564877584236544)]} 0 5
   [junit4]   2> 3987680 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[391]} 0 8
   [junit4]   2> 3987684 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[420 (1440564877593673728)]} 0 1
   [junit4]   2> 3987690 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[420 (1440564877597868032)]} 0 0
   [junit4]   2> 3987692 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[420 (1440564877597868032)]} 0 5
   [junit4]   2> 3987725 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[374 (-1440564877637713920)]} 0 1
   [junit4]   2> 3987731 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[374 (-1440564877640859648)]} 0 0
   [junit4]   2> 3987732 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[374 (-1440564877640859648)]} 0 4
   [junit4]   2> 3987736 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[421 (1440564877648199680)]} 0 1
   [junit4]   2> 3987745 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[421 (1440564877655539712)]} 0 1
   [junit4]   2> 3987746 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[421 (1440564877655539712)]} 0 4
   [junit4]   2> 3987748 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[421]} 0 9
   [junit4]   2> 3987783 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[296 (-1440564877698531328)]} 0 1
   [junit4]   2> 3987791 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[296 (-1440564877704822784)]} 0 0
   [junit4]   2> 3987792 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[296 (-1440564877704822784)]} 0 3
   [junit4]   2> 3987793 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[296]} 0 7
   [junit4]   2> 3987797 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[422 (1440564877712162816)]} 0 1
   [junit4]   2> 3987803 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[422 (1440564877716357120)]} 0 1
   [junit4]   2> 3987805 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[422 (1440564877716357120)]} 0 5
   [junit4]   2> 3987837 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[396 (-1440564877755154432)]} 0 0
   [junit4]   2> 3987851 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[396 (-1440564877765640192)]} 0 1
   [junit4]   2> 3987852 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[396 (-1440564877765640192)]} 0 5
   [junit4]   2> 3987853 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[396]} 0 12
   [junit4]   2> 3987861 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[423 (1440564877779271680)]} 0 1
   [junit4]   2> 3987871 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[423 (1440564877787660288)]} 0 1
   [junit4]   2> 3987872 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[423 (1440564877787660288)]} 0 5
   [junit4]   2> 3987910 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[207 (-1440564877830651904)]} 0 2
   [junit4]   2> 3987921 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[207 (-1440564877840089088)]} 0 0
   [junit4]   2> 3987922 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[207 (-1440564877840089088)]} 0 4
   [junit4]   2> 3987923 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[207]} 0 8
   [junit4]   2> 3987927 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[424 (1440564877849526272)]} 0 1
   [junit4]   2> 3987938 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[424 (1440564877857914880)]} 0 0
   [junit4]   2> 3987940 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[424 (1440564877857914880)]} 0 5
   [junit4]   2> 3987941 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[424]} 0 9
   [junit4]   2> 3987978 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[425 (1440564877898809344)]} 0 5
   [junit4]   2> 3987996 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[425 (1440564877912440832)]} 0 1
   [junit4]   2> 3987997 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[425 (1440564877912440832)]} 0 10
   [junit4]   2> 3987998 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[425]} 0 15
   [junit4]   2> 3988033 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[426 (1440564877959626752)]} 0 1
   [junit4]   2> 3988040 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[426 (1440564877964869632)]} 0 1
   [junit4]   2> 3988041 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[426 (1440564877964869632)]} 0 4
   [junit4]   2> 3988074 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[427 (1440564878002618368)]} 0 1
   [junit4]   2> 3988086 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[427 (1440564878012055552)]} 0 1
   [junit4]   2> 3988088 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[427 (1440564878012055552)]} 0 6
   [junit4]   2> 3988089 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[427]} 0 10
   [junit4]   2> 3988123 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[428 (1440564878053998592)]} 0 1
   [junit4]   2> 3988135 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[428 (1440564878063435776)]} 0 1
   [junit4]   2> 3988136 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[428 (1440564878063435776)]} 0 5
   [junit4]   2> 3988139 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[428]} 0 11
   [junit4]   2> 3988172 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[321 (-1440564878106427392)]} 0 0
   [junit4]   2> 3988183 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[321 (-1440564878113767424)]} 0 0
   [junit4]   2> 3988185 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[321 (-1440564878113767424)]} 0 7
   [junit4]   2> 3988191 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[429 (1440564878125301760)]} 0 1
   [junit4]   2> 3988201 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[429 (1440564878132641792)]} 0 1
   [junit4]   2> 3988202 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[429 (1440564878132641792)]} 0 5
   [junit4]   2> 3988203 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[429]} 0 9
   [junit4]   2> 3988236 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[430 (1440564878172487680)]} 0 1
   [junit4]   2> 3988256 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[430 (1440564878184022016)]} 0 1
   [junit4]   2> 3988257 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[430 (1440564878184022016)]} 0 11
   [junit4]   2> 3988258 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[430]} 0 18
   [junit4]   2> 3988291 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[428 (-1440564878230159360)]} 0 1
   [junit4]   2> 3988299 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[428 (-1440564878236450816)]} 0 0
   [junit4]   2> 3988301 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[428 (-1440564878236450816)]} 0 5
   [junit4]   2> 3988302 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[428]} 0 8
   [junit4]   2> 3988307 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[431 (1440564878245888000)]} 0 2
   [junit4]   2> 3988329 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[431 (1440564878267908096)]} 0 1
   [junit4]   2> 3988330 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[431 (1440564878267908096)]} 0 5
   [junit4]   2> 3988331 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[431]} 0 17
   [junit4]   2> 3988365 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[407 (-1440564878307753984)]} 0 1
   [junit4]   2> 3988373 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[407 (-1440564878315094016)]} 0 0
   [junit4]   2> 3988375 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[407 (-1440564878315094016)]} 0 4
   [junit4]   2> 3988376 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[407]} 0 7
   [junit4]   2> 3988384 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[432 (1440564878327676928)]} 0 3
   [junit4]   2> 3988398 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[432 (1440564878340259840)]} 0 1
   [junit4]   2> 3988400 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[432 (1440564878340259840)]} 0 5
   [junit4]   2> 3988401 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[432]} 0 9
   [junit4]   2> 3988434 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[376 (-1440564878381154304)]} 0 0
   [junit4]   2> 3988448 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[376 (-1440564878392688640)]} 0 0
   [junit4]   2> 3988452 T8026 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[376 (-1440564878392688640)]} 0 8
   [junit4]   2> 3988453 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[376]} 0 13
   [junit4]   2> 3988459 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[433 (1440564878406320128)]} 0 1
   [junit4]   2> 3988469 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[433 (1440564878413660160)]} 0 1
   [junit4]   2> 3988470 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[433 (1440564878413660160)]} 0 5
   [junit4]   2> 3988471 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[433]} 0 9
   [junit4]   2> 3988511 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[434 (1440564878458748928)]} 0 5
   [junit4]   2> 3988522 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[434 (1440564878469234688)]} 0 1
   [junit4]   2> 3988523 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[434 (1440564878469234688)]} 0 5
   [junit4]   2> 3988557 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[435 (1440564878509080576)]} 0 1
   [junit4]   2> 3988568 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[435 (1440564878517469184)]} 0 1
   [junit4]   2> 3988570 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[435 (1440564878517469184)]} 0 6
   [junit4]   2> 3988572 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[435]} 0 11
   [junit4]   2> 3988605 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[371 (-1440564878560460800)]} 0 0
   [junit4]   2> 3988616 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[371 (-1440564878568849408)]} 0 0
   [junit4]   2> 3988618 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[371 (-1440564878568849408)]} 0 5
   [junit4]   2> 3988620 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[371]} 0 10
   [junit4]   2> 3988624 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[436 (1440564878580383744)]} 0 1
   [junit4]   2> 3988632 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[436 (1440564878584578048)]} 0 1
   [junit4]   2> 3988634 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[436 (1440564878584578048)]} 0 6
   [junit4]   2> 3988667 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[250 (-1440564878625472512)]} 0 0
   [junit4]   2> 3988674 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[250 (-1440564878629666816)]} 0 0
   [junit4]   2> 3988676 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[250 (-1440564878629666816)]} 0 5
   [junit4]   2> 3988681 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[437 (1440564878639104000)]} 0 1
   [junit4]   2> 3988690 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[437 (1440564878646444032)]} 0 0
   [junit4]   2> 3988692 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[437 (1440564878646444032)]} 0 5
   [junit4]   2> 3988693 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[437]} 0 9
   [junit4]   2> 3988727 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[438 (1440564878688387072)]} 0 1
   [junit4]   2> 3988741 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[438 (1440564878693629952)]} 0 1
   [junit4]   2> 3988742 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[438 (1440564878693629952)]} 0 11
   [junit4]   2> 3988777 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[439 (1440564878739767296)]} 0 1
   [junit4]   2> 3988787 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[439 (1440564878748155904)]} 0 0
   [junit4]   2> 3988789 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[439 (1440564878748155904)]} 0 5
   [junit4]   2> 3988791 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[439]} 0 10
   [junit4]   2> 3988825 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[432 (-1440564878791147520)]} 0 1
   [junit4]   2> 3988839 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[432 (-1440564878802681856)]} 0 0
   [junit4]   2> 3988841 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[432 (-1440564878802681856)]} 0 5
   [junit4]   2> 3988842 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[432]} 0 13
   [junit4]   2> 3988846 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[440 (1440564878813167616)]} 0 0
   [junit4]   2> 3988854 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[440 (1440564878819459072)]} 0 0
   [junit4]   2> 3988855 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[440 (1440564878819459072)]} 0 3
   [junit4]   2> 3988856 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[440]} 0 6
   [junit4]   2> 3988893 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[441 (1440564878861402112)]} 0 2
   [junit4]   2> 3988908 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[441 (1440564878871887872)]} 0 1
   [junit4]   2> 3988910 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[441 (1440564878871887872)]} 0 8
   [junit4]   2> 3988945 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[202 (-1440564878916976640)]} 0 0
   [junit4]   2> 3988954 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[202 (-1440564878923268096)]} 0 1
   [junit4]   2> 3988955 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[202 (-1440564878923268096)]} 0 4
   [junit4]   2> 3988956 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[202]} 0 8
   [junit4]   2> 3988959 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[442 (1440564878931656704)]} 0 1
   [junit4]   2> 3988967 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[442 (1440564878936899584)]} 0 0
   [junit4]   2> 3988968 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[442 (1440564878936899584)]} 0 4
   [junit4]   2> 3988969 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[442]} 0 7
   [junit4]   2> 3989009 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[292 (-1440564878983036928)]} 0 2
   [junit4]   2> 3989017 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[292 (-1440564878990376960)]} 0 0
   [junit4]   2> 3989019 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[292 (-1440564878990376960)]} 0 4
   [junit4]   2> 3989019 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[292]} 0 7
   [junit4]   2> 3989030 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[443 (1440564879005057024)]} 0 1
   [junit4]   2> 3989047 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[443 (1440564879019737088)]} 0 0
   [junit4]   2> 3989049 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[443 (1440564879019737088)]} 0 6
   [junit4]   2> 3989051 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[443]} 0 16
   [junit4]   2> 3989083 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[201 (-1440564879061680128)]} 0 0
   [junit4]   2> 3989094 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[201 (-1440564879071117312)]} 0 0
   [junit4]   2> 3989098 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[201 (-1440564879071117312)]} 0 6
   [junit4]   2> 3989106 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[444 (1440564879084748800)]} 0 1
   [junit4]   2> 3989114 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[444 (1440564879091040256)]} 0 0
   [junit4]   2> 3989116 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[444 (1440564879091040256)]} 0 5
   [junit4]   2> 3989150 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[364 (-1440564879130886144)]} 0 1
   [junit4]   2> 3989162 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[364 (-1440564879141371904)]} 0 0
   [junit4]   2> 3989163 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[364 (-1440564879141371904)]} 0 4
   [junit4]   2> 3989164 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[364]} 0 8
   [junit4]   2> 3989168 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[445 (1440564879149760512)]} 0 1
   [junit4]   2> 3989174 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[445 (1440564879153954816)]} 0 0
   [junit4]   2> 3989176 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[445 (1440564879153954816)]} 0 5
   [junit4]   2> 3989209 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[275 (-1440564879193800704)]} 0 0
   [junit4]   2> 3989226 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[275 (-1440564879197995008)]} 0 0
   [junit4]   2> 3989227 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[275 (-1440564879197995008)]} 0 14
   [junit4]   2> 3989231 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[446 (1440564879215820800)]} 0 1
   [junit4]   2> 3989248 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[446 (1440564879230500864)]} 0 0
   [junit4]   2> 3989251 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[446 (1440564879230500864)]} 0 8
   [junit4]   2> 3989253 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[446]} 0 19
   [junit4]   2> 3989286 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[320 (-1440564879274541056)]} 0 0
   [junit4]   2> 3989300 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[320 (-1440564879286075392)]} 0 1
   [junit4]   2> 3989301 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[320 (-1440564879286075392)]} 0 4
   [junit4]   2> 3989302 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[320]} 0 8
   [junit4]   2> 3989310 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[447 (1440564879298658304)]} 0 3
   [junit4]   2> 3989317 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[447 (1440564879303901184)]} 0 1
   [junit4]   2> 3989318 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[447 (1440564879303901184)]} 0 5
   [junit4]   2> 3989354 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[448 (1440564879344795648)]} 0 1
   [junit4]   2> 3989368 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[448 (1440564879354232832)]} 0 1
   [junit4]   2> 3989369 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[448 (1440564879354232832)]} 0 7
   [junit4]   2> 3989370 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[448]} 0 11
   [junit4]   2> 3989404 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[200 (-1440564879398273024)]} 0 1
   [junit4]   2> 3989422 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[200 (-1440564879411904512)]} 0 0
   [junit4]   2> 3989423 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[200 (-1440564879411904512)]} 0 6
   [junit4]   2> 3989425 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[200]} 0 16
   [junit4]   2> 3989434 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[449 (1440564879428681728)]} 0 1
   [junit4]   2> 3989442 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[449 (1440564879433924608)]} 0 1
   [junit4]   2> 3989443 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[449 (1440564879433924608)]} 0 6
   [junit4]   2> 3989477 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[450 (1440564879473770496)]} 0 1
   [junit4]   2> 3989490 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[450 (1440564879485304832)]} 0 0
   [junit4]   2> 3989492 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[450 (1440564879485304832)]} 0 6
   [junit4]   2> 3989493 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[450]} 0 10
   [junit4]   2> 3989526 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[451 (1440564879526199296)]} 0 0
   [junit4]   2> 3989535 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[451 (1440564879532490752)]} 0 0
   [junit4]   2> 3989537 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[451 (1440564879532490752)]} 0 5
   [junit4]   2> 3989538 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[451]} 0 8
   [junit4]   2> 3989579 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[452 (1440564879581773824)]} 0 0
   [junit4]   2> 3989587 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[452 (1440564879588065280)]} 0 0
   [junit4]   2> 3989588 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[452 (1440564879588065280)]} 0 3
   [junit4]   2> 3989589 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[452]} 0 7
   [junit4]   2> 3989622 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[453 (1440564879626862592)]} 0 1
   [junit4]   2> 3989638 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[453 (1440564879638396928)]} 0 1
   [junit4]   2> 3989639 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[453 (1440564879638396928)]} 0 6
   [junit4]   2> 3989640 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[453]} 0 11
   [junit4]   2> 3989678 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[454 (1440564879683485696)]} 0 4
   [junit4]   2> 3989699 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[454 (1440564879703408640)]} 0 1
   [junit4]   2> 3989700 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[454 (1440564879703408640)]} 0 5
   [junit4]   2> 3989701 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[454]} 0 14
   [junit4]   2> 3989735 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[455 (1440564879744303104)]} 0 1
   [junit4]   2> 3989749 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[455 (1440564879751643136)]} 0 3
   [junit4]   2> 3989751 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[455 (1440564879751643136)]} 0 10
   [junit4]   2> 3989752 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[455]} 0 14
   [junit4]   2> 3989785 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[221 (-1440564879797780480)]} 0 0
   [junit4]   2> 3989795 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[221 (-1440564879805120512)]} 0 1
   [junit4]   2> 3989797 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[221 (-1440564879805120512)]} 0 5
   [junit4]   2> 3989798 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[221]} 0 9
   [junit4]   2> 3989803 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[456 (1440564879815606272)]} 0 1
   [junit4]   2> 3989816 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[456 (1440564879820849152)]} 0 2
   [junit4]   2> 3989818 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[456 (1440564879820849152)]} 0 11
   [junit4]   2> 3989850 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[306 (-1440564879865937920)]} 0 0
   [junit4]   2> 3989859 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[306 (-1440564879872229376)]} 0 1
   [junit4]   2> 3989860 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[306 (-1440564879872229376)]} 0 4
   [junit4]   2> 3989861 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[306]} 0 8
   [junit4]   2> 3989865 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[457 (1440564879880617984)]} 0 1
   [junit4]   2> 3989891 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[457 (1440564879898443776)]} 0 1
   [junit4]   2> 3989892 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[457 (1440564879898443776)]} 0 14
   [junit4]   2> 3989893 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[457]} 0 21
   [junit4]   2> 3989926 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[458 (1440564879945629696)]} 0 0
   [junit4]   2> 3989935 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[458 (1440564879952969728)]} 0 0
   [junit4]   2> 3989936 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[458 (1440564879952969728)]} 0 4
   [junit4]   2> 3989938 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[458]} 0 8
   [junit4]   2> 3989975 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[459 (1440564879997009920)]} 0 0
   [junit4]   2> 3989984 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[459 (1440564880004349952)]} 0 0
   [junit4]   2> 3989986 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[459 (1440564880004349952)]} 0 5
   [junit4]   2> 3989987 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[459]} 0 8
   [junit4]   2> 3990023 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[460 (1440564880046292992)]} 0 1
   [junit4]   2> 3990032 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[460 (1440564880053633024)]} 0 0
   [junit4]   2> 3990034 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[460 (1440564880053633024)]} 0 5
   [junit4]   2> 3990035 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[460]} 0 9
   [junit4]   2> 3990069 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[461 (1440564880094527488)]} 0 1
   [junit4]   2> 3990086 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[461 (1440564880102916096)]} 0 1
   [junit4]   2> 3990087 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[461 (1440564880102916096)]} 0 12
   [junit4]   2> 3990119 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[425 (-1440564880148004864)]} 0 0
   [junit4]   2> 3990128 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[425 (-1440564880154296320)]} 0 1
   [junit4]   2> 3990129 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[425 (-1440564880154296320)]} 0 4
   [junit4]   2> 3990131 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[425]} 0 8
   [junit4]   2> 3990134 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[462 (1440564880163733504)]} 0 1
   [junit4]   2> 3990149 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[462 (1440564880175267840)]} 0 1
   [junit4]   2> 3990150 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[462 (1440564880175267840)]} 0 5
   [junit4]   2> 3990151 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[462]} 0 12
   [junit4]   2> 3990185 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[463 (1440564880216162304)]} 0 1
   [junit4]   2> 3990196 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[463 (1440564880225599488)]} 0 0
   [junit4]   2> 3990198 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[463 (1440564880225599488)]} 0 5
   [junit4]   2> 3990199 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[463]} 0 9
   [junit4]   2> 3990233 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[464 (1440564880266493952)]} 0 1
   [junit4]   2> 3990243 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[464 (1440564880274882560)]} 0 1
   [junit4]   2> 3990244 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[464 (1440564880274882560)]} 0 5
   [junit4]   2> 3990246 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[464]} 0 9
   [junit4]   2> 3990278 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[436 (-1440564880314728448)]} 0 0
   [junit4]   2> 3990285 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[436 (-1440564880318922752)]} 0 0
   [junit4]   2> 3990287 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[436 (-1440564880318922752)]} 0 5
   [junit4]   2> 3990295 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[465 (1440564880332554240)]} 0 1
   [junit4]   2> 3990310 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[465 (1440564880341991424)]} 0 2
   [junit4]   2> 3990312 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[465 (1440564880341991424)]} 0 8
   [junit4]   2> 3990314 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[465]} 0 14
   [junit4]   2> 3990350 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[196 (-1440564880390225920)]} 0 0
   [junit4]   2> 3990365 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[196 (-1440564880403857408)]} 0 0
   [junit4]   2> 3990367 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[196 (-1440564880403857408)]} 0 5
   [junit4]   2> 3990368 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[196]} 0 10
   [junit4]   2> 3990372 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[466 (1440564880413294592)]} 0 1
   [junit4]   2> 3990380 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[466 (1440564880419586048)]} 0 0
   [junit4]   2> 3990382 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[466 (1440564880419586048)]} 0 4
   [junit4]   2> 3990383 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[466]} 0 7
   [junit4]   2> 3990416 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[467 (1440564880458383360)]} 0 1
   [junit4]   2> 3990432 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[467 (1440564880468869120)]} 0 1
   [junit4]   2> 3990433 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[467 (1440564880468869120)]} 0 8
   [junit4]   2> 3990434 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[467]} 0 11
   [junit4]   2> 3990468 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[367 (-1440564880512909312)]} 0 2
   [junit4]   2> 3990480 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[367 (-1440564880523395072)]} 0 1
   [junit4]   2> 3990481 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[367 (-1440564880523395072)]} 0 4
   [junit4]   2> 3990487 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[468 (1440564880532832256)]} 0 0
   [junit4]   2> 3990497 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[468 (1440564880540172288)]} 0 1
   [junit4]   2> 3990498 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[468 (1440564880540172288)]} 0 5
   [junit4]   2> 3990534 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[370 (-1440564880582115328)]} 0 2
   [junit4]   2> 3990546 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[370 (-1440564880592601088)]} 0 0
   [junit4]   2> 3990551 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[370 (-1440564880592601088)]} 0 8
   [junit4]   2> 3990552 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[370]} 0 11
   [junit4]   2> 3990561 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[469 (1440564880611475456)]} 0 1
   [junit4]   2> 3990570 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[469 (1440564880615669760)]} 0 1
   [junit4]   2> 3990571 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[469 (1440564880615669760)]} 0 6
   [junit4]   2> 3990608 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[358 (-1440564880660758528)]} 0 1
   [junit4]   2> 3990618 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[358 (-1440564880668098560)]} 0 0
   [junit4]   2> 3990620 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[358 (-1440564880668098560)]} 0 5
   [junit4]   2> 3990627 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[470 (1440564880680681472)]} 0 0
   [junit4]   2> 3990637 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[470 (1440564880686972928)]} 0 1
   [junit4]   2> 3990638 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[470 (1440564880686972928)]} 0 5
   [junit4]   2> 3990640 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[470]} 0 9
   [junit4]   2> 3990673 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[456 (-1440564880727867392)]} 0 0
   [junit4]   2> 3990685 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[456 (-1440564880739401728)]} 0 0
   [junit4]   2> 3990687 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[456 (-1440564880739401728)]} 0 4
   [junit4]   2> 3990691 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[471 (1440564880747790336)]} 0 1
   [junit4]   2> 3990703 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[471 (1440564880757227520)]} 0 1
   [junit4]   2> 3990704 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[471 (1440564880757227520)]} 0 4
   [junit4]   2> 3990706 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[471]} 0 12
   [junit4]   2> 3990742 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[251 (-1440564880801267712)]} 0 1
   [junit4]   2> 3990760 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[251 (-1440564880811753472)]} 0 1
   [junit4]   2> 3990763 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[251 (-1440564880811753472)]} 0 12
   [junit4]   2> 3990765 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[251]} 0 17
   [junit4]   2> 3990769 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[472 (1440564880829579264)]} 0 1
   [junit4]   2> 3990779 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[472 (1440564880835870720)]} 0 1
   [junit4]   2> 3990780 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[472 (1440564880835870720)]} 0 5
   [junit4]   2> 3990781 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[472]} 0 8
   [junit4]   2> 3990815 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[136 (-1440564880877813760)]} 0 0
   [junit4]   2> 3990832 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[136 (-1440564880892493824)]} 0 0
   [junit4]   2> 3990834 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[136 (-1440564880892493824)]} 0 5
   [junit4]   2> 3990835 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[136]} 0 10
   [junit4]   2> 3990839 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[473 (1440564880902979584)]} 0 0
   [junit4]   2> 3990847 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[473 (1440564880907173888)]} 0 1
   [junit4]   2> 3990849 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[473 (1440564880907173888)]} 0 6
   [junit4]   2> 3990882 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[392 (-1440564880948068352)]} 0 1
   [junit4]   2> 3990893 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[392 (-1440564880955408384)]} 0 0
   [junit4]   2> 3990894 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[392 (-1440564880955408384)]} 0 6
   [junit4]   2> 3990899 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[474 (1440564880965894144)]} 0 1
   [junit4]   2> 3990908 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[474 (1440564880972185600)]} 0 1
   [junit4]   2> 3990909 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[474 (1440564880972185600)]} 0 5
   [junit4]   2> 3990911 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[474]} 0 9
   [junit4]   2> 3990945 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[475 (1440564881013080064)]} 0 1
   [junit4]   2> 3990958 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[475 (1440564881023565824)]} 0 1
   [junit4]   2> 3990960 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[475 (1440564881023565824)]} 0 6
   [junit4]   2> 3990961 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[475]} 0 12
   [junit4]   2> 3990995 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[476 (1440564881065508864)]} 0 1
   [junit4]   2> 3991009 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[476 (1440564881072848896)]} 0 3
   [junit4]   2> 3991011 T8026 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[476 (1440564881072848896)]} 0 10
   [junit4]   2> 3991012 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[476]} 0 14
   [junit4]   2> 3991045 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[336 (-1440564881118986240)]} 0 0
   [junit4]   2> 3991056 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[336 (-1440564881125277696)]} 0 1
   [junit4]   2> 3991058 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[336 (-1440564881125277696)]} 0 7
   [junit4]   2> 3991059 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[336]} 0 11
   [junit4]   2> 3991066 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[477 (1440564881141006336)]} 0 1
   [junit4]   2> 3991099 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[477 (1440564881165123584)]} 0 1
   [junit4]   2> 3991104 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[477 (1440564881165123584)]} 0 15
   [junit4]   2> 3991111 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[477]} 0 33
   [junit4]   2> 3991145 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[124 (-1440564881223843840)]} 0 0
   [junit4]   2> 3991160 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[124 (-1440564881236426752)]} 0 0
   [junit4]   2> 3991162 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[124 (-1440564881236426752)]} 0 5
   [junit4]   2> 3991166 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[124]} 0 11
   [junit4]   2> 3991178 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[478 (1440564881257398272)]} 0 1
   [junit4]   2> 3991194 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[478 (1440564881269981184)]} 0 1
   [junit4]   2> 3991200 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[478 (1440564881269981184)]} 0 11
   [junit4]   2> 3991206 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[478]} 0 24
   [junit4]   2> 3991241 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[479 (1440564881324507136)]} 0 1
   [junit4]   2> 3991254 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[479 (1440564881332895744)]} 0 1
   [junit4]   2> 3991255 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[479 (1440564881332895744)]} 0 6
   [junit4]   2> 3991256 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[479]} 0 12
   [junit4]   2> 3991289 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[454 (-1440564881374838784)]} 0 0
   [junit4]   2> 3991302 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[454 (-1440564881386373120)]} 0 0
   [junit4]   2> 3991321 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[454 (-1440564881386373120)]} 0 21
   [junit4]   2> 3991323 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[454]} 0 26
   [junit4]   2> 3991335 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[480 (1440564881422024704)]} 0 1
   [junit4]   2> 3991348 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[480 (1440564881432510464)]} 0 1
   [junit4]   2> 3991350 T8024 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[480 (1440564881432510464)]} 0 6
   [junit4]   2> 3991355 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[480]} 0 14
   [junit4]   2> 3991393 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[481 (1440564881482842112)]} 0 1
   [junit4]   2> 3991405 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[481 (1440564881492279296)]} 0 1
   [junit4]   2> 3991407 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[481 (1440564881492279296)]} 0 6
   [junit4]   2> 3991412 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[481]} 0 14
   [junit4]   2> 3991446 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[404 (-1440564881539465216)]} 0 0
   [junit4]   2> 3991463 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[404 (-1440564881554145280)]} 0 0
   [junit4]   2> 3991465 T8025 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[404 (-1440564881554145280)]} 0 5
   [junit4]   2> 3991466 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[404]} 0 9
   [junit4]   2> 3991470 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[482 (1440564881563582464)]} 0 1
   [junit4]   2> 3991486 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[482 (1440564881568825344)]} 0 10
   [junit4]   2> 3991489 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[482 (1440564881568825344)]} 0 16
   [junit4]   2> 3991526 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[483 (1440564881620205568)]} 0 3
   [junit4]   2> 3991536 T8044 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[483 (1440564881629642752)]} 0 0
   [junit4]   2> 3991538 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[483 (1440564881629642752)]} 0 6
   [junit4]   2> 3991586 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[325 (-1440564881685217280)]} 0 1
   [junit4]   2> 3991592 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[325 (-1440564881689411584)]} 0 0
   [junit4]   2> 3991594 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[325 (-1440564881689411584)]} 0 5
   [junit4]   2> 3991599 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[484 (1440564881698848768)]} 0 2
   [junit4]   2> 3991616 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[484 (1440564881706188800)]} 0 1
   [junit4]   2> 3991620 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[484 (1440564881706188800)]} 0 15
   [junit4]   2> 3991659 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[394 (-1440564881761763328)]} 0 1
   [junit4]   2> 3991666 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[394 (-1440564881765957632)]} 0 1
   [junit4]   2> 3991667 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[394 (-1440564881765957632)]} 0 5
   [junit4]   2> 3991671 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[485 (1440564881775394816)]} 0 0
   [junit4]   2> 3991682 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[485 (1440564881782734848)]} 0 1
   [junit4]   2> 3991692 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[485 (1440564881782734848)]} 0 14
   [junit4]   2> 3991696 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[485]} 0 21
   [junit4]   2> 3991736 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[486 (1440564881842503680)]} 0 1
   [junit4]   2> 3991751 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[486 (1440564881855086592)]} 0 1
   [junit4]   2> 3991752 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[486 (1440564881855086592)]} 0 6
   [junit4]   2> 3991753 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[486]} 0 14
   [junit4]   2> 3991792 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[487 (1440564881901223936)]} 0 1
   [junit4]   2> 3991799 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[487 (1440564881906466816)]} 0 1
   [junit4]   2> 3991800 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[487 (1440564881906466816)]} 0 4
   [junit4]   2> 3991833 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[259 (-1440564881945264128)]} 0 0
   [junit4]   2> 3991844 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[259 (-1440564881953652736)]} 0 1
   [junit4]   2> 3991855 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[259 (-1440564881953652736)]} 0 15
   [junit4]   2> 3991865 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[259]} 0 28
   [junit4]   2> 3991881 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[488 (1440564881994547200)]} 0 1
   [junit4]   2> 3991889 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[488 (1440564881999790080)]} 0 1
   [junit4]   2> 3991896 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[488 (1440564881999790080)]} 0 11
   [junit4]   2> 3991930 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[359 (-1440564882046976000)]} 0 0
   [junit4]   2> 3991940 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[359 (-1440564882054316032)]} 0 1
   [junit4]   2> 3991942 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[359 (-1440564882054316032)]} 0 5
   [junit4]   2> 3991949 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[359]} 0 15
   [junit4]   2> 3991958 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[489 (1440564882076336128)]} 0 1
   [junit4]   2> 3991970 T8039 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[489 (1440564882084724736)]} 0 0
   [junit4]   2> 3991972 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[489 (1440564882084724736)]} 0 6
   [junit4]   2> 3992015 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[299 (-1440564882135056384)]} 0 1
   [junit4]   2> 3992032 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[299 (-1440564882149736448)]} 0 1
   [junit4]   2> 3992043 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[299 (-1440564882149736448)]} 0 15
   [junit4]   2> 3992044 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[299]} 0 19
   [junit4]   2> 3992049 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[490 (1440564882170707968)]} 0 1
   [junit4]   2> 3992058 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[490 (1440564882178048000)]} 0 1
   [junit4]   2> 3992059 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[490 (1440564882178048000)]} 0 4
   [junit4]   2> 3992060 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[490]} 0 8
   [junit4]   2> 3992094 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[266 (-1440564882217893888)]} 0 1
   [junit4]   2> 3992104 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[266 (-1440564882225233920)]} 0 1
   [junit4]   2> 3992105 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[266 (-1440564882225233920)]} 0 5
   [junit4]   2> 3992107 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[266]} 0 10
   [junit4]   2> 3992112 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[491 (1440564882236768256)]} 0 1
   [junit4]   2> 3992125 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[491 (1440564882247254016)]} 0 1
   [junit4]   2> 3992127 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[491 (1440564882247254016)]} 0 6
   [junit4]   2> 3992163 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[244 (-1440564882291294208)]} 0 1
   [junit4]   2> 3992174 T8005 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[244 (-1440564882299682816)]} 0 1
   [junit4]   2> 3992176 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[244 (-1440564882299682816)]} 0 6
   [junit4]   2> 3992177 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[244]} 0 9
   [junit4]   2> 3992181 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[492 (1440564882310168576)]} 0 1
   [junit4]   2> 3992191 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[492 (1440564882317508608)]} 0 0
   [junit4]   2> 3992192 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[492 (1440564882317508608)]} 0 4
   [junit4]   2> 3992194 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[492]} 0 9
   [junit4]   2> 3992232 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[493 (1440564882362597376)]} 0 1
   [junit4]   2> 3992249 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[493 (1440564882374131712)]} 0 2
   [junit4]   2> 3992253 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[493 (1440564882374131712)]} 0 11
   [junit4]   2> 3992257 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[493]} 0 19
   [junit4]   2> 3992295 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[494 (1440564882428657664)]} 0 4
   [junit4]   2> 3992304 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[494 (1440564882435997696)]} 0 1
   [junit4]   2> 3992305 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[494 (1440564882435997696)]} 0 4
   [junit4]   2> 3992306 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[494]} 0 7
   [junit4]   2> 3992340 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[265 (-1440564882476892160)]} 0 0
   [junit4]   2> 3992347 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[265 (-1440564882481086464)]} 0 0
   [junit4]   2> 3992348 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[265 (-1440564882481086464)]} 0 4
   [junit4]   2> 3992352 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[495 (1440564882488426496)]} 0 1
   [junit4]   2> 3992358 T8041 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[495 (1440564882492620800)]} 0 0
   [junit4]   2> 3992359 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[495 (1440564882492620800)]} 0 4
   [junit4]   2> 3992393 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[496 (1440564882531418112)]} 0 1
   [junit4]   2> 3992407 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[496 (1440564882537709568)]} 0 1
   [junit4]   2> 3992408 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[496 (1440564882537709568)]} 0 10
   [junit4]   2> 3992409 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[496]} 0 13
   [junit4]   2> 3992444 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[497 (1440564882584895488)]} 0 1
   [junit4]   2> 3992464 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[497 (1440564882599575552)]} 0 2
   [junit4]   2> 3992468 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[497 (1440564882599575552)]} 0 11
   [junit4]   2> 3992473 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[497]} 0 21
   [junit4]   2> 3992513 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[498 (1440564882657247232)]} 0 1
   [junit4]   2> 3992523 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[498 (1440564882665635840)]} 0 0
   [junit4]   2> 3992525 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[498 (1440564882665635840)]} 0 5
   [junit4]   2> 3992526 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[498]} 0 9
   [junit4]   2> 3992559 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[499 (1440564882706530304)]} 0 0
   [junit4]   2> 3992569 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[499 (1440564882713870336)]} 0 0
   [junit4]   2> 3992571 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[499 (1440564882713870336)]} 0 5
   [junit4]   2> 3992572 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[499]} 0 9
   [junit4]   2> 3992605 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[494 (-1440564882754764800)]} 0 0
   [junit4]   2> 3992614 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[494 (-1440564882762104832)]} 0 0
   [junit4]   2> 3992616 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[494 (-1440564882762104832)]} 0 5
   [junit4]   2> 3992617 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[494]} 0 8
   [junit4]   2> 3992620 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[500 (1440564882770493440)]} 0 1
   [junit4]   2> 3992632 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[500 (1440564882776784896)]} 0 1
   [junit4]   2> 3992634 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[500 (1440564882776784896)]} 0 9
   [junit4]   2> 3992635 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[500]} 0 12
   [junit4]   2> 3992669 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[393 (-1440564882821873664)]} 0 1
   [junit4]   2> 3992675 T8042 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[393 (-1440564882825019392)]} 0 0
   [junit4]   2> 3992677 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[393 (-1440564882825019392)]} 0 5
   [junit4]   2> 3992680 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[501 (1440564882833408000)]} 0 1
   [junit4]   2> 3992692 T8004 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[501 (1440564882842845184)]} 0 0
   [junit4]   2> 3992693 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[501 (1440564882842845184)]} 0 4
   [junit4]   2> 3992694 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[501]} 0 11
   [junit4]   2> 3992727 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[442 (-1440564882882691072)]} 0 1
   [junit4]   2> 3992739 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[442 (-1440564882892128256)]} 0 1
   [junit4]   2> 3992740 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[442 (-1440564882892128256)]} 0 4
   [junit4]   2> 3992742 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[442]} 0 11
   [junit4]   2> 3992747 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[502 (1440564882902614016)]} 0 1
   [junit4]   2> 3992756 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[502 (1440564882909954048)]} 0 0
   [junit4]   2> 3992758 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[502 (1440564882909954048)]} 0 5
   [junit4]   2> 3992760 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[502]} 0 10
   [junit4]   2> 3992792 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[503 (1440564882950848512)]} 0 0
   [junit4]   2> 3992807 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[503 (1440564882963431424)]} 0 0
   [junit4]   2> 3992808 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[503 (1440564882963431424)]} 0 4
   [junit4]   2> 3992809 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[503]} 0 7
   [junit4]   2> 3992844 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[504 (1440564883004325888)]} 0 1
   [junit4]   2> 3992858 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[504 (1440564883015860224)]} 0 1
   [junit4]   2> 3992859 T8028 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[504 (1440564883015860224)]} 0 5
   [junit4]   2> 3992861 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[504]} 0 10
   [junit4]   2> 3992894 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[463 (-1440564883057803264)]} 0 0
   [junit4]   2> 3992903 T8006 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[463 (-1440564883064094720)]} 0 0
   [junit4]   2> 3992904 T8026 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[463 (-1440564883064094720)]} 0 4
   [junit4]   2> 3992906 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[463]} 0 8
   [junit4]   2> 3992911 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[505 (1440564883073531904)]} 0 2
   [junit4]   2> 3992929 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[505 (1440564883086114816)]} 0 1
   [junit4]   2> 3992930 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[505 (1440564883086114816)]} 0 9
   [junit4]   2> 3992931 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[505]} 0 13
   [junit4]   2> 3992965 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[506 (1440564883131203584)]} 0 1
   [junit4]   2> 3992975 T8008 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[506 (1440564883139592192)]} 0 1
   [junit4]   2> 3992977 T8107 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[506 (1440564883139592192)]} 0 6
   [junit4]   2> 3992978 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[506]} 0 9
   [junit4]   2> 3993014 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[507 (1440564883182583808)]} 0 1
   [junit4]   2> 3993024 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[507 (1440564883189923840)]} 0 1
   [junit4]   2> 3993025 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[507 (1440564883189923840)]} 0 5
   [junit4]   2> 3993027 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[507]} 0 9
   [junit4]   2> 3993060 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[426 (-1440564883231866880)]} 0 1
   [junit4]   2> 3993067 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[426 (-1440564883236061184)]} 0 0
   [junit4]   2> 3993071 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[426 (-1440564883236061184)]} 0 7
   [junit4]   2> 3993076 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[508 (1440564883248644096)]} 0 1
   [junit4]   2> 3993085 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[508 (1440564883254935552)]} 0 0
   [junit4]   2> 3993087 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[508 (1440564883254935552)]} 0 5
   [junit4]   2> 3993088 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[508]} 0 9
   [junit4]   2> 3993121 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[509 (1440564883295830016)]} 0 1
   [junit4]   2> 3993132 T8072 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[509 (1440564883302121472)]} 0 2
   [junit4]   2> 3993133 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[509 (1440564883302121472)]} 0 6
   [junit4]   2> 3993135 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[509]} 0 9
   [junit4]   2> 3993168 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[510 (1440564883344064512)]} 0 1
   [junit4]   2> 3993182 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[510 (1440564883351404544)]} 0 1
   [junit4]   2> 3993183 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[510 (1440564883351404544)]} 0 9
   [junit4]   2> 3993184 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[510]} 0 13
   [junit4]   2> 3993218 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[511 (1440564883396493312)]} 0 1
   [junit4]   2> 3993227 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[511 (1440564883403833344)]} 0 0
   [junit4]   2> 3993229 T8042 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[511 (1440564883403833344)]} 0 5
   [junit4]   2> 3993230 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[511]} 0 9
   [junit4]   2> 3993263 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[512 (1440564883443679232)]} 0 1
   [junit4]   2> 3993271 T8043 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[512 (1440564883448922112)]} 0 0
   [junit4]   2> 3993274 T8023 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[512 (1440564883448922112)]} 0 7
   [junit4]   2> 3993307 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[513 (1440564883490865152)]} 0 0
   [junit4]   2> 3993324 T8077 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[513 (1440564883501350912)]} 0 1
   [junit4]   2> 3993325 T8040 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[513 (1440564883501350912)]} 0 10
   [junit4]   2> 3993326 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[513]} 0 15
   [junit4]   2> 3993360 T8007 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[514 (1440564883546439680)]} 0 0
   [junit4]   2> 3993370 T8003 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[514 (1440564883553779712)]} 0 0
   [junit4]   2> 3993372 T8027 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[514 (1440564883553779712)]} 0 5
   [junit4]   2> 3993373 T8026 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[514]} 0 9
   [junit4]   2> 3993408 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[224 (-1440564883596771328)]} 0 1
   [junit4]   2> 3993420 T8074 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[224 (-1440564883606208512)]} 0 0
   [junit4]   2> 3993422 T8039 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[224 (-1440564883606208512)]} 0 5
   [junit4]   2> 3993423 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[224]} 0 9
   [junit4]   2> 3993427 T8004 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[515 (1440564883615645696)]} 0 1
   [junit4]   2> 3993489 T8073 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[515 (1440564883655491584)]} 0 1
   [junit4]   2> 3993491 T8041 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[515 (1440564883655491584)]} 0 28
   [junit4]   2> 3993492 T8107 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[515]} 0 62
   [junit4]   2> 3993529 T8006 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[231 (-1440564883723649024)]} 0 0
   [junit4]   2> 3993538 T8007 C1870 P51251 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[231 (-1440564883729940480)]} 0 0
   [junit4]   2> 3993540 T8023 C1871 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_1_replica1] webapp=/qb path=/update params={update.distrib=TOLEADER&wt=javabin&version=2} {delete=[231 (-1440564883729940480)]} 0 5
   [junit4]   2> 3993541 T8028 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[231]} 0 8
   [junit4]   2> 3993576 T8005 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[516 (1440564883771883520)]} 0 1
   [junit4]   2> 3993599 T8076 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[516 (1440564883792855040)]} 0 1
   [junit4]   2> 3993600 T8044 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[516 (1440564883792855040)]} 0 6
   [junit4]   2> 3993601 T8025 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[516]} 0 9
   [junit4]   2> 3993664 T8003 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[517 (1440564883865206784)]} 0 0
   [junit4]   2> 3993674 T8075 C1877 P51265 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51258/qb/collection1/&update.distrib=FROMLEADER&wt=javabin&version=2} {add=[517 (1440564883872546816)]} 0 0
   [junit4]   2> 3993675 T8043 C1876 P51258 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=TOLEADER&wt=javabin&version=2} {add=[517 (1440564883872546816)]} 0 4
   [junit4]   2> 3993677 T8027 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {add=[517]} 0 9
   [junit4]   2> 3993712 T8008 C1874 P51251 oasup.LogUpdateProcessor.finish [collection1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[361 (-1440564883915538432)]} 0 0
   [junit4]   2> 3993719 T8040 C1869 P51258 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica2] webapp=/qb path=/update params={distrib.from=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/&update.distrib=FROMLEADER&wt=javabin&version=2} {delete=[361 (-1440564883920781312)]} 0 0
   [junit4]   2> 3993721 T8024 C1875 P51255 oasup.LogUpdateProcessor.finish [collection1_shard1_0_replica1] webapp=/qb path=/update params={wt=javabin&version=2} {delete=[361 (-1440564883920781312)]} 0 5
   [junit4]   2> 3993736 T7990 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3993739 T7990 oasc.ShardSplitTest.checkDocCountsAndShardStates ShardSplitTest waited for 0 ms for shard state to be set to active
   [junit4]   2> 3993740 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3993774 T8107 C1875 P51255 oasc.SolrCore.execute [collection1_shard1_0_replica1] webapp=/qb path=/select params={fl=id,_version_&q=*:*&distrib=false&wt=javabin&version=2&rows=1000} hits=47 status=0 QTime=1 
   [junit4]   2> 3993778 T7990 oasc.ShardSplitTest.checkSubShardConsistency Shard: shard1_0 Replica: http://127.0.0.1:51255/qb/collection1_shard1_0_replica1/ has 47 docs
   [junit4]   2> 3993779 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3993811 T8039 C1869 P51258 oasc.SolrCore.execute [collection1_shard1_0_replica2] webapp=/qb path=/select params={fl=id,_version_&q=*:*&distrib=false&wt=javabin&version=2&rows=1000} hits=47 status=0 QTime=1 
   [junit4]   2> 3993812 T7990 oasc.ShardSplitTest.checkSubShardConsistency Shard: shard1_0 Replica: http://127.0.0.1:51258/qb/collection1_shard1_0_replica2/ has 47 docs
   [junit4]   2> 3993813 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3993838 T8023 C1871 P51255 oasc.SolrCore.execute [collection1_shard1_1_replica1] webapp=/qb path=/select params={fl=id,_version_&q=*:*&distrib=false&wt=javabin&version=2&rows=1000} hits=81 status=0 QTime=0 
   [junit4]   2> 3993843 T7990 oasc.ShardSplitTest.checkSubShardConsistency Shard: shard1_1 Replica: http://127.0.0.1:51255/qb/collection1_shard1_1_replica1/ has 81 docs
   [junit4]   2> 3993844 T7990 oascsi.HttpClientUtil.createClient Creating new http client, config:maxConnections=128&maxConnectionsPerHost=32&followRedirects=false
   [junit4]   2> 3993864 T8006 C1870 P51251 oasc.SolrCore.execute [collection1_shard1_1_replica2] webapp=/qb path=/select params={fl=id,_version_&q=*:*&distrib=false&wt=javabin&version=2&rows=1000} hits=80 status=0 QTime=1 
   [junit4]   2> 3993876 T7990 oasc.ShardSplitTest.checkSubShardConsistency Shard: shard1_1 Replica: http://127.0.0.1:51251/qb/collection1_shard1_1_replica2/ has 80 docs
   [junit4]   2> 3993889 T7990 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 51251
   [junit4]   2> 3993889 T7990 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1171924669
   [junit4]   2> 3994831 T8013 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3994835 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1_shard1_1_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3994836 T8013 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1_1 is already registered
   [junit4]   2> 3994852 T8013 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3994853 T8013 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 3994865 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3994866 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3994865 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3994865 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3994866 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3995919 T7990 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 3995920 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3995923 T7990 oasc.ZkController.publish publishing core=collection1_shard1_1_replica2 state=down
   [junit4]   2> 3995923 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3995927 T7990 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@7a780607
   [junit4]   2> 3995936 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=2,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=417,adds=417,deletesById=208,deletesByQuery=0,errors=0,cumulative_adds=518,cumulative_deletesById=208,cumulative_deletesByQuery=1,cumulative_errors=0}
   [junit4]   2> 3995936 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 3995937 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 3995937 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2>  C1874_STATE=coll:control_collection core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb, leader=true}
   [junit4]   2> 3995944 T7990 C1874 P51251 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@65e17bbd lockFactory=org.apache.lucene.store.NativeFSLockFactory@7c5b2df5),segFN=segments_2,generation=2}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@65e17bbd lockFactory=org.apache.lucene.store.NativeFSLockFactory@7c5b2df5),segFN=segments_3,generation=3}
   [junit4]   2> 3995944 T7990 C1874 P51251 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 3
   [junit4]   2> 3995948 T7990 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 3995949 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 3995950 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data;done=false>>]
   [junit4]   2> 3995951 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data
   [junit4]   2> 3995951 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data/index [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data/index;done=false>>]
   [junit4]   2> 3995952 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/control/data/index
   [junit4]   2> 3995952 T7990 oasc.SolrCore.close [collection1_shard1_1_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@1562478b
   [junit4]   2> 3995965 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=2,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=43,adds=43,deletesById=29,deletesByQuery=0,errors=0,cumulative_adds=62,cumulative_deletesById=35,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 3995966 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 3995967 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 3995967 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2>  C1870_STATE=coll:collection1 core:collection1_shard1_1_replica2 props:{state=down, core=collection1_shard1_1_replica2, node_name=127.0.0.1:51251_qb, base_url=http://127.0.0.1:51251/qb}
   [junit4]   2> 3995973 T7990 C1870 P51251 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@22ec99d4 lockFactory=org.apache.lucene.store.NativeFSLockFactory@22d86580),segFN=segments_5,generation=5}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@22ec99d4 lockFactory=org.apache.lucene.store.NativeFSLockFactory@22d86580),segFN=segments_6,generation=6}
   [junit4]   2> 3995974 T7990 C1870 P51251 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 6
   [junit4]   2> 3995979 T7990 oasc.SolrCore.closeSearcher [collection1_shard1_1_replica2] Closing main searcher on request.
   [junit4]   2> 3995980 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 3995980 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data;done=false>>]
   [junit4]   2> 3995981 T7990 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data
   [junit4]   2> 3995981 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index.20130714202156725 [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index.20130714202156725;done=false>>]
   [junit4]   2> 3995982 T7990 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-controljetty-1373829666446//collection1_shard1_1_replica2/data/index.20130714202156725
   [junit4]   2> 3995988 T8013 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035301003689987-127.0.0.1:51251_qb-n_0000000000) am no longer a leader.
   [junit4]   2> 3996000 T8012 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3996003 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3996006 T8064 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3996007 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3996007 T8048 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3996008 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3996010 T8032 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 3996015 T8081 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3996022 T7990 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/qb,null}
   [junit4]   2> 3996024 T8032 oasc.Overseer.start Overseer (id=90035301003689990-127.0.0.1:51255_qb-n_0000000001) starting
   [junit4]   2> 3996037 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 5)
   [junit4]   2> 3996038 T8125 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 3996042 T8032 oascc.ZkStateReader$3.process Updating live nodes... (4)
   [junit4]   2> 3996044 T8124 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 3996049 T8124 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3996051 T8124 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3996061 T8124 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node8",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1_shard1_1_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51251_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   2> 3996071 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3996071 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3996072 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3996072 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3996079 T7990 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 51255
   [junit4]   2> 3996081 T7990 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=939315043
   [junit4]   2> 3997302 T8012 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 3997306 T8012 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 3997597 T8124 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3997601 T8124 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3997601 T8124 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 3997612 T8124 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1_shard1_0_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3997612 T8124 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1_0 is already registered
   [junit4]   2> 3997622 T8124 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1_shard1_1_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3997622 T8124 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1_1 is already registered
   [junit4]   2> 3997633 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3997634 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3997634 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3997635 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3999099 T7990 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 3999100 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3999104 T7990 oasc.ZkController.publish publishing core=collection1_shard1_1_replica1 state=down
   [junit4]   2> 3999107 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3999115 T7990 oasc.ZkController.publish publishing core=collection1_shard1_0_replica1 state=down
   [junit4]   2> 3999117 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 3999120 T7990 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@31a4adfc
   [junit4]   2> 3999131 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=5,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=3,adds=3,deletesById=2,deletesByQuery=0,errors=0,cumulative_adds=176,cumulative_deletesById=47,cumulative_deletesByQuery=1,cumulative_errors=0}
   [junit4]   2> 3999132 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 3999133 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 3999133 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2>  C1873_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3999137 T7990 C1873 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_4,generation=4}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@33973381 lockFactory=org.apache.lucene.store.NativeFSLockFactory@29f7c2d0),segFN=segments_5,generation=5}
   [junit4]   2> 3999138 T7990 C1873 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 5
   [junit4]   2> 3999144 T7990 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 3999145 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 3999146 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1/index [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1/index;done=false>>]
   [junit4]   2> 3999147 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1/index
   [junit4]   2> 3999148 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1 [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1;done=false>>]
   [junit4]   2> 3999149 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty1
   [junit4]   2> 3999149 T7990 oasc.SolrCore.close [collection1_shard1_1_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@54aa02ed
   [junit4]   2> 3999158 T8124 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3999189 T8124 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3999198 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=3,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=43,adds=42,deletesById=29,deletesByQuery=0,errors=0,cumulative_adds=91,cumulative_deletesById=59,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 3999199 T8124 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node6",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":null,
   [junit4]   2> 	  "core":"collection1_shard1_1_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3999199 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 3999200 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 3999200 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> 3999209 T8124 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node5",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_0",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":null,
   [junit4]   2> 	  "core":"collection1_shard1_0_replica1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51255_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51255/qb"}
   [junit4]   2> 3999218 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3999218 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3999219 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3999218 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 4)
   [junit4]   2> ASYNC  NEW_CORE C1878 name=collection1_shard1_1_replica1 org.apache.solr.core.SolrCore@54aa02ed url=http://127.0.0.1:51255/qb/collection1_shard1_1_replica1 node=127.0.0.1:51255_qb C1878_STATE=coll:collection1 core:collection1_shard1_1_replica1 props:{state=down, core=collection1_shard1_1_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3999239 T7990 C1878 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=3
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_4,generation=4}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7b67d479 lockFactory=org.apache.lucene.store.NativeFSLockFactory@7d825b02),segFN=segments_5,generation=5}
   [junit4]   2> 3999240 T7990 C1878 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 5
   [junit4]   2> 3999246 T7990 oasc.SolrCore.closeSearcher [collection1_shard1_1_replica1] Closing main searcher on request.
   [junit4]   2> 3999247 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 3999248 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data/index;done=false>>]
   [junit4]   2> 3999248 T7990 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data/index
   [junit4]   2> 3999249 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data;done=false>>]
   [junit4]   2> 3999250 T7990 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_1_replica1/data
   [junit4]   2> 3999251 T7990 oasc.SolrCore.close [collection1_shard1_0_replica1]  CLOSING SolrCore org.apache.solr.core.SolrCore@5c1f6bce
   [junit4]   2> 3999271 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=3,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=56,adds=56,deletesById=33,deletesByQuery=0,errors=0,cumulative_adds=98,cumulative_deletesById=63,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 3999275 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 3999275 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 3999276 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> ASYNC  NEW_CORE C1879 name=collection1_shard1_0_replica1 org.apache.solr.core.SolrCore@5c1f6bce url=http://127.0.0.1:51255/qb/collection1_shard1_0_replica1 node=127.0.0.1:51255_qb C1879_STATE=coll:collection1 core:collection1_shard1_0_replica1 props:{state=down, core=collection1_shard1_0_replica1, node_name=127.0.0.1:51255_qb, base_url=http://127.0.0.1:51255/qb, leader=true}
   [junit4]   2> 3999282 T7990 C1879 P51255 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=3
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_4,generation=4}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@5d597373 lockFactory=org.apache.lucene.store.NativeFSLockFactory@28222b03),segFN=segments_5,generation=5}
   [junit4]   2> 3999283 T7990 C1879 P51255 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 5
   [junit4]   2> 3999288 T7990 oasc.SolrCore.closeSearcher [collection1_shard1_0_replica1] Closing main searcher on request.
   [junit4]   2> 3999289 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 3999289 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data/index [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data/index;done=false>>]
   [junit4]   2> 3999290 T7990 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data/index
   [junit4]   2> 3999291 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data;done=false>>]
   [junit4]   2> 3999292 T7990 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty1-1373829672033//collection1_shard1_0_replica1/data
   [junit4]   2> 3999296 T8124 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035301003689990-127.0.0.1:51255_qb-n_0000000001) am no longer a leader.
   [junit4]   2> 3999308 T8032 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3999309 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 4)
   [junit4]   2> 3999312 T8064 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 3999312 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3999317 T8048 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 3999320 T8081 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 3999325 T8064 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 3999328 T8048 oasc.Overseer.start Overseer (id=90035301003689992-127.0.0.1:51258_qb-n_0000000002) starting
   [junit4]   2> 3999333 T8064 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 3999334 T8064 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 3999334 T8064 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 3999334 T8064 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:51261/qb/collection1/
   [junit4]   2> 3999335 T8064 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 3999336 T8064 oasc.SyncStrategy.syncToMe http://127.0.0.1:51261/qb/collection1/ has no replicas
   [junit4]   2> 3999336 T8064 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:51261/qb/collection1/ shard1
   [junit4]   2> 3999337 T8064 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard1
   [junit4]   2> 3999346 T8127 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 3999348 T7990 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/qb,null}
   [junit4]   2> 3999349 T8048 oascc.ZkStateReader$3.process Updating live nodes... (3)
   [junit4]   2> 3999354 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3999355 T8126 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 3999363 T8048 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1_0
   [junit4]   2> 3999366 T8126 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 3999370 T8048 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 3999371 T8048 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 3999371 T8048 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 3999371 T8048 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:51258/qb/collection1_shard1_0_replica2/
   [junit4]   2> 3999372 T8048 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 3999372 T8048 oasc.SyncStrategy.syncToMe http://127.0.0.1:51258/qb/collection1_shard1_0_replica2/ has no replicas
   [junit4]   2> 3999372 T8048 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:51258/qb/collection1_shard1_0_replica2/ shard1_0
   [junit4]   2> 3999373 T8048 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard1_0
   [junit4]   2> 3999408 T7990 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 51258
   [junit4]   2> 3999409 T7990 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=848682637
   [junit4]   2> 3999412 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3999412 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 3999414 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 4000610 T8032 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 4000610 T8032 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 4000957 T8126 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 4000960 T8126 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 4000960 T8126 oasc.Overseer$ClusterStateUpdater.updateState shard=shard2 is already registered
   [junit4]   2> 4000972 T8126 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1_shard1_0_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 4000973 T8126 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1_0 is already registered
   [junit4]   2> 4000983 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 4000985 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 4000987 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 4002422 T7990 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 4002423 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 4002426 T7990 oasc.ZkController.publish publishing core=collection1_shard1_0_replica2 state=down
   [junit4]   2> 4002427 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 4002430 T7990 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@143a584e
   [junit4]   2> 4002439 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=4,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=87,adds=87,deletesById=41,deletesByQuery=0,errors=0,cumulative_adds=247,cumulative_deletesById=103,cumulative_deletesByQuery=1,cumulative_errors=0}
   [junit4]   2> 4002440 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 4002441 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 4002442 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> ASYNC  NEW_CORE C1880 name=collection1 org.apache.solr.core.SolrCore@143a584e url=http://127.0.0.1:51258/qb/collection1 node=127.0.0.1:51258_qb C1880_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 4002449 T7990 C1880 P51258 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@7caf62cd lockFactory=org.apache.lucene.store.NativeFSLockFactory@65cdf0fd),segFN=segments_4,generation=4}
   [junit4]   2> 4002450 T7990 C1880 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 4
   [junit4]   2> 4002460 T7990 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 4002460 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 4002461 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2/index [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2/index;done=false>>]
   [junit4]   2> 4002461 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2/index
   [junit4]   2> 4002462 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2 [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2;done=false>>]
   [junit4]   2> 4002462 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty2
   [junit4]   2> 4002462 T7990 oasc.SolrCore.close [collection1_shard1_0_replica2]  CLOSING SolrCore org.apache.solr.core.SolrCore@200e6d1f
   [junit4]   2> 4002473 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=2,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=56,adds=56,deletesById=31,deletesByQuery=0,errors=0,cumulative_adds=76,cumulative_deletesById=38,cumulative_deletesByQuery=0,cumulative_errors=0}
   [junit4]   2> 4002474 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 4002474 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 4002475 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> ASYNC  NEW_CORE C1881 name=collection1_shard1_0_replica2 org.apache.solr.core.SolrCore@200e6d1f url=http://127.0.0.1:51258/qb/collection1_shard1_0_replica2 node=127.0.0.1:51258_qb C1881_STATE=coll:collection1 core:collection1_shard1_0_replica2 props:{state=down, core=collection1_shard1_0_replica2, node_name=127.0.0.1:51258_qb, base_url=http://127.0.0.1:51258/qb, leader=true}
   [junit4]   2> 4002480 T7990 C1881 P51258 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2f254955 lockFactory=org.apache.lucene.store.NativeFSLockFactory@23e45ac7),segFN=segments_5,generation=5}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@2f254955 lockFactory=org.apache.lucene.store.NativeFSLockFactory@23e45ac7),segFN=segments_6,generation=6}
   [junit4]   2> 4002480 T7990 C1881 P51258 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 6
   [junit4]   2> 4002485 T7990 oasc.SolrCore.closeSearcher [collection1_shard1_0_replica2] Closing main searcher on request.
   [junit4]   2> 4002486 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 4002487 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data;done=false>>]
   [junit4]   2> 4002488 T7990 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data
   [junit4]   2> 4002488 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index.20130714202156647 [CachedDir<<refCount=0;path=/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index.20130714202156647;done=false>>]
   [junit4]   2> 4002489 T7990 oasc.CachingDirectoryFactory.close Closing directory: /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-jetty2-1373829678140//collection1_shard1_0_replica2/data/index.20130714202156647
   [junit4]   2> 4002491 T8126 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035301003689992-127.0.0.1:51258_qb-n_0000000002) am no longer a leader.
   [junit4]   2> 4002497 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 4002501 T8081 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 4002513 T8081 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 4002514 T8064 oascc.ZkStateReader$3.process Updating live nodes... (2)
   [junit4]   2> 4002515 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 4002519 T8081 oasc.ShardLeaderElectionContext.shouldIBeLeader Checking if I should try and be the leader.
   [junit4]   2> 4002519 T8081 oasc.ShardLeaderElectionContext.shouldIBeLeader My last published State was Active, it's okay to be the leader.
   [junit4]   2> 4002519 T8081 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 4002520 T8081 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:51265/qb/collection1/
   [junit4]   2> 4002520 T8081 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 4002520 T8081 oasc.SyncStrategy.syncToMe http://127.0.0.1:51265/qb/collection1/ has no replicas
   [junit4]   2> 4002521 T8081 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:51265/qb/collection1/ shard2
   [junit4]   2> 4002521 T8081 oascc.SolrZkClient.makePath makePath: /collections/collection1/leaders/shard2
   [junit4]   2> 4002521 T8064 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 4002529 T7990 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/qb,null}
   [junit4]   2> 4002534 T8064 oasc.Overseer.start Overseer (id=90035301003689994-127.0.0.1:51261_qb-n_0000000003) starting
   [junit4]   2> 4002550 T8129 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 4002554 T8128 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 4002560 T8128 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 4002563 T8128 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node2",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 4002574 T8128 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node7",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1_0",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1_shard1_0_replica2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51258_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51258/qb"}
   [junit4]   2> 4002584 T7990 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 51261
   [junit4]   2> 4002585 T7990 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1975274070
   [junit4]   2> 4002613 T8128 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=null message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "node_name":"127.0.0.1:51261_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51261/qb"}
   [junit4]   2> 4002614 T8128 oasc.Overseer$ClusterStateUpdater.updateState shard=shard1 is already registered
   [junit4]   2> 4002623 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 4002624 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 4003590 T7990 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 4003591 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 4003594 T7990 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@6a1f4eb5
   [junit4]   2> 4003600 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=3,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=4,adds=4,deletesById=2,deletesByQuery=0,errors=0,cumulative_adds=176,cumulative_deletesById=47,cumulative_deletesByQuery=1,cumulative_errors=0}
   [junit4]   2> 4003600 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 4003601 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 4003601 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> ASYNC  NEW_CORE C1882 name=collection1 org.apache.solr.core.SolrCore@6a1f4eb5 url=http://127.0.0.1:51261/qb/collection1 node=127.0.0.1:51261_qb C1882_STATE=coll:collection1 core:collection1 props:{state=down, core=collection1, node_name=127.0.0.1:51261_qb, base_url=http://127.0.0.1:51261/qb, leader=true}
   [junit4]   2> 4003605 T7990 C1882 P51261 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@4cf6277a lockFactory=org.apache.lucene.store.NativeFSLockFactory@518a7bdf),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@4cf6277a lockFactory=org.apache.lucene.store.NativeFSLockFactory@518a7bdf),segFN=segments_4,generation=4}
   [junit4]   2> 4003613 T7990 C1882 P51261 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 4
   [junit4]   2> 4003618 T7990 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 4003619 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 4003619 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3/index [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3/index;done=false>>]
   [junit4]   2> 4003619 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3/index
   [junit4]   2> 4003620 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3 [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3;done=false>>]
   [junit4]   2> 4003620 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty3
   [junit4]   2> 4003622 T8128 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035301003689994-127.0.0.1:51261_qb-n_0000000003) am no longer a leader.
   [junit4]   2> 4003627 T8064 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 4003628 T8064 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 4003628 T8064 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 4003629 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 2)
   [junit4]   2> 4003634 T8081 oascc.ZkStateReader$3.process Updating live nodes... (1)
   [junit4]   2> 4003638 T8081 oascc.SolrZkClient.makePath makePath: /overseer_elect/leader
   [junit4]   2> 4003644 T7990 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/qb,null}
   [junit4]   2> 4003651 T8081 oasc.Overseer.start Overseer (id=90035301003689996-127.0.0.1:51265_qb-n_0000000004) starting
   [junit4]   2> 4003668 T8131 oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 4003674 T8130 oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 4003680 T8130 oascc.ZkStateReader.updateClusterState Updating cloud state from ZooKeeper... 
   [junit4]   2> 4003684 T8130 oasc.Overseer$ClusterStateUpdater.updateState Update state numShards=2 message={
   [junit4]   2> 	  "operation":"state",
   [junit4]   2> 	  "core_node_name":"core_node3",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard_state":"active",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "shard_range":null,
   [junit4]   2> 	  "node_name":"127.0.0.1:51261_qb",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:51261/qb"}
   [junit4]   2> 4003697 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeDataChanged path:/clusterstate.json, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 4003700 T7990 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 51265
   [junit4]   2> 4003702 T7990 oasc.CoreContainer.shutdown Shutting down CoreContainer instance=1583629964
   [junit4]   2> 4003805 T8048 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 4003805 T8048 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 3)
   [junit4]   2> 4003806 T8048 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 4004706 T7990 oasc.ZkController.publish publishing core=collection1 state=down
   [junit4]   2> 4004709 T7990 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 4004714 T7990 oasc.SolrCore.close [collection1]  CLOSING SolrCore org.apache.solr.core.SolrCore@33d7fe66
   [junit4]   2> 4004727 T7990 oasu.DirectUpdateHandler2.close closing DirectUpdateHandler2{commits=3,autocommits=0,soft autocommits=0,optimizes=0,rollbacks=0,expungeDeletes=0,docsPending=87,adds=87,deletesById=41,deletesByQuery=0,errors=0,cumulative_adds=247,cumulative_deletesById=103,cumulative_deletesByQuery=1,cumulative_errors=0}
   [junit4]   2> 4004728 T7990 oasu.SolrCoreState.decrefSolrCoreState Closing SolrCoreState
   [junit4]   2> 4004729 T7990 oasu.DefaultSolrCoreState.closeIndexWriter SolrCoreState ref count has reached 0 - closing IndexWriter
   [junit4]   2> 4004729 T7990 oasu.DefaultSolrCoreState.closeIndexWriter closing IndexWriter with IndexWriterCloser
   [junit4]   2> ASYNC  NEW_CORE C1883 name=collection1 org.apache.solr.core.SolrCore@33d7fe66 url=http://127.0.0.1:51265/qb/collection1 node=127.0.0.1:51265_qb C1883_STATE=coll:collection1 core:collection1 props:{state=active, core=collection1, node_name=127.0.0.1:51265_qb, base_url=http://127.0.0.1:51265/qb, leader=true}
   [junit4]   2> 4004734 T7990 C1883 P51265 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=2
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191fa213 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1b45ecd0),segFN=segments_3,generation=3}
   [junit4]   2> 		commit{dir=MockDirWrapper(org.apache.lucene.store.RAMDirectory@191fa213 lockFactory=org.apache.lucene.store.NativeFSLockFactory@1b45ecd0),segFN=segments_4,generation=4}
   [junit4]   2> 4004735 T7990 C1883 P51265 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 4
   [junit4]   2> 4004740 T7990 oasc.SolrCore.closeSearcher [collection1] Closing main searcher on request.
   [junit4]   2> 4004741 T7990 oasc.CachingDirectoryFactory.close Closing MockDirectoryFactory - 2 directories currently being tracked
   [junit4]   2> 4004741 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4/index [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4/index;done=false>>]
   [junit4]   2> 4004742 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4/index
   [junit4]   2> 4004742 T7990 oasc.CachingDirectoryFactory.closeCacheValue looking to close ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4 [CachedDir<<refCount=0;path=./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4;done=false>>]
   [junit4]   2> 4004743 T7990 oasc.CachingDirectoryFactory.close Closing directory: ./org.apache.solr.cloud.ShardSplitTest-1373829666185/jetty4
   [junit4]   2> 4004744 T8130 oasc.Overseer$ClusterStateUpdater.amILeader According to ZK I (id=90035301003689996-127.0.0.1:51265_qb-n_0000000004) am no longer a leader.
   [junit4]   2> 4004749 T8081 oascc.ZkStateReader$2.process A cluster state change: WatchedEvent state:SyncConnected type:NodeChildrenChanged path:/live_nodes, has occurred - updating... (live nodes size: 1)
   [junit4]   2> 4004765 T7990 oejsh.ContextHandler.doStop stopped o.e.j.s.ServletContextHandler{/qb,null}
   [junit4]   2> 4004856 T7990 oas.SolrTestCaseJ4.tearDown ###Ending testDistribSearch
   [junit4]   2> 4004860 T7990 oasc.ZkTestServer.send4LetterWord connecting to 127.0.0.1:51248 51248
   [junit4]   2> 4004890 T7990 oasc.ChaosMonkey.monkeyLog monkey: stop shard! 51251
   [junit4]   2> !!!! WARNING: best effort to remove /Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build/solr-core/test/J0/./org.apache.solr.cloud.ShardSplitTest-1373829666185 FAILED !!!!!
   [junit4]   2> 4004891 T7990 oas.SolrTestCaseJ4.tearDown ###Ending testDistribSearch
   [junit4]   2> 4004891 T7990 oasc.ZkTestServer.send4LetterWord connecting to 127.0.0.1:51248 51248
   [junit4]   1>            "core":"collection1",
   [junit4]   1>            "node_name":"127.0.0.1:51251_qb",
   [junit4]   1>            "base_url":"http://127.0.0.1:51251/qb"}
   [junit4]   1>   /solr/overseer_elect (2)
   [junit4]   1>    /solr/overseer_elect/election (5)
   [junit4]   1>     /solr/overseer_elect/election/90035301003689987-127.0.0.1:51251_qb-n_0000000000 (0)
   [junit4]   1>     /solr/overseer_elect/election/90035301003689992-127.0.0.1:51258_qb-n_0000000002 (0)
   [junit4]   1>     /solr/overseer_elect/election/90035301003689990-127.0.0.1:51255_qb-n_0000000001 (0)
   [junit4]   1>     /solr/overseer_elect/election/90035301003689996-127.0.0.1:51265_qb-n_0000000004 (0)
   [junit4]   1>     /solr/overseer_elect/election/90035301003689994-127.0.0.1:51261_qb-n_0000000003 (0)
   [junit4]   1>    /solr/overseer_elect/leader (0)
   [junit4]   1>    DATA:
   [junit4]   1>        {"id":"90035301003689987-127.0.0.1:51251_qb-n_0000000000"}
   [junit4]   1>  /zookeeper (1)
   [junit4]   1>  DATA:
   [junit4]   1>      
   [junit4]   1> 
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=ShardSplitTest -Dtests.method=testDistribSearch -Dtests.seed=30BF22C5938FEB5 -Dtests.slow=true -Dtests.locale=in -Dtests.timezone=Africa/Douala -Dtests.file.encoding=US-ASCII
   [junit4] FAILURE 75.9s | ShardSplitTest.testDistribSearch <<<
   [junit4]    > Throwable #1: java.lang.AssertionError: shard1_1 is not consistent expected:<81> but was:<80>
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([30BF22C5938FEB5:82ED7C342E679E89]:0)
   [junit4]    > 	at org.apache.solr.cloud.ShardSplitTest.checkSubShardConsistency(ShardSplitTest.java:247)
   [junit4]    > 	at org.apache.solr.cloud.ShardSplitTest.checkDocCountsAndShardStates(ShardSplitTest.java:207)
   [junit4]    > 	at org.apache.solr.cloud.ShardSplitTest.doTest(ShardSplitTest.java:173)
   [junit4]    > 	at org.apache.solr.BaseDistributedSearchTestCase.testDistribSearch(BaseDistributedSearchTestCase.java:835)
   [junit4]    > 	at java.lang.Thread.run(Thread.java:724)
   [junit4]   2> 4004907 T7990 oas.SolrTestCaseJ4.deleteCore ###deleteCore
   [junit4]   2> 75938 T7989 ccr.ThreadLeakControl.checkThreadLeaks WARNING Will linger awaiting termination of 2 leaked thread(s).
   [junit4]   2> 4006051 T8081 oascc.ZkStateReader$2.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> 4006052 T8081 oascc.ZkStateReader$3.process WARN ZooKeeper watch triggered, but Solr cannot talk to ZK
   [junit4]   2> NOTE: test params are: codec=Lucene41, sim=DefaultSimilarity, locale=in, timezone=Africa/Douala
   [junit4]   2> NOTE: Mac OS X 10.8.4 x86_64/Oracle Corporation 1.7.0_25 (64-bit)/cpus=2,threads=1,free=192910072,total=405471232
   [junit4]   2> NOTE: All tests run in this JVM: [BadComponentTest, DistributedQueryElevationComponentTest, RAMDirectoryFactoryTest, TestReloadAndDeleteDocs, IndexSchemaTest, SpellCheckComponentTest, PeerSyncTest, ZkControllerTest, TestStressVersions, DistributedTermsComponentTest, ChaosMonkeyNothingIsSafeTest, FieldMutatingUpdateProcessorTest, LegacyHTMLStripCharFilterTest, TestSolrXMLSerializer, SOLR749Test, TestSolrQueryParser, TestFieldTypeCollectionResource, PreAnalyzedUpdateProcessorTest, TestSolrJ, TestCloudManagedSchema, TestQuerySenderNoQuery, HdfsBasicDistributedZk2Test, ResponseLogComponentTest, TestQuerySenderListener, TestWriterPerf, CoreAdminCreateDiscoverTest, SearchHandlerTest, TestPostingsSolrHighlighter, TestClassNameShortening, CollectionsAPIDistributedZkTest, SoftAutoCommitTest, DistanceFunctionTest, TestConfig, DocumentAnalysisRequestHandlerTest, ZkNodePropsTest, TestCSVLoader, SuggesterTSTTest, TestGroupingSearch, BasicDistributedZkTest, TestSweetSpotSimilarityFactory, TestManagedSchema, OpenCloseCoreStressTest, AddSchemaFieldsUpdateProcessorFactoryTest, SuggesterTest, TestSolrCoreProperties, TestAddFieldRealTimeGet, TestLazyCores, SpellPossibilityIteratorTest, TestSort, TestHashPartitioner, SolrCoreTest, FullSolrCloudDistribCmdsTest, TestTrie, DebugComponentTest, TestPartialUpdateDeduplication, TestReplicationHandler, HdfsSyncSliceTest, ClusterStateTest, SchemaVersionSpecificBehaviorTest, CSVRequestHandlerTest, ConvertedLegacyTest, AlternateDirectoryTest, PolyFieldTest, PluginInfoTest, TestDynamicFieldResource, TestFieldTypeResource, TestValueSourceCache, QueryResultKeyTest, BasicFunctionalityTest, AnalysisAfterCoreReloadTest, TestSchemaResource, TestSuggestSpellingConverter, ChaosMonkeySafeLeaderTest, TestJmxMonitoredMap, HardAutoCommitTest, TestSolrXmlPersistence, TestSchemaSimilarityResource, TestDefaultSimilarityFactory, TestInfoStreamLogging, TestPhraseSuggestions, AliasIntegrationTest, TestRandomDVFaceting, TestFuzzyAnalyzedSuggestions, TestQueryUtils, TestSchemaNameResource, UpdateRequestProcessorFactoryTest, TestPHPSerializedResponseWriter, ScriptEngineTest, TestSystemIdResolver, TestQueryTypes, MultiTermTest, TestFastWriter, SignatureUpdateProcessorFactoryTest, TestSolrQueryParserResource, PrimUtilsTest, HdfsBasicDistributedZkTest, TestIBSimilarityFactory, ParsingFieldUpdateProcessorsTest, TestPerFieldSimilarity, URLClassifyProcessorTest, CoreContainerCoreInitFailuresTest, TestZkChroot, ClusterStateUpdateTest, TestManagedSchemaFieldResource, SimplePostToolTest, TestCodecSupport, JsonLoaderTest, TestJmxIntegration, DocValuesTest, RequestHandlersTest, DefaultValueUpdateProcessorTest, DocValuesMultiTest, QueryParsingTest, ExternalFileFieldSortTest, TestCharFilters, TestSchemaVersionResource, TestXIncludeConfig, StatsComponentTest, CacheHeaderTest, PathHierarchyTokenizerFactoryTest, TestPropInjectDefaults, IndexBasedSpellCheckerTest, SystemInfoHandlerTest, TestCloudManagedSchemaAddField, TestSerializedLuceneMatchVersion, CurrencyFieldXmlFileTest, LukeRequestHandlerTest, TestExtendedDismaxParser, TestSolrQueryParserDefaultOperatorResource, TestNumberUtils, CurrencyFieldOpenExchangeTest, TestStressReorder, TestShardHandlerFactory, CopyFieldTest, RegexBoostProcessorTest, MinimalSchemaTest, QueryElevationComponentTest, CircularListTest, DocumentBuilderTest, TestAtomicUpdateErrorCases, DateFieldTest, TermVectorComponentDistributedTest, CoreAdminHandlerTest, TestComponentsName, BlockCacheTest, SolrIndexSplitterTest, SolrCmdDistributorTest, TestCopyFieldCollectionResource, TestFastOutputStream, BasicDistributedZk2Test, UnloadDistributedZkTest, SyncSliceTest, OverseerTest, LeaderElectionIntegrationTest, ShardRoutingTest, BasicZkTest, RecoveryZkTest, TestRandomFaceting, LeaderElectionTest, ZkSolrClientTest, ZkCLITest, ShardRoutingCustomTest, SliceStateUpdateTest, TestDistributedSearch, AutoCommitTest, TestDistributedGrouping, TestFaceting, TestRecovery, DistributedSpellCheckComponentTest, TestRealTimeGet, TestJoin, TestMultiCoreConfBootstrap, TestReload, TestRangeQuery, TestCoreContainer, SimpleFacetsTest, TestSolr4Spatial, BadIndexSchemaTest, TestFiltering, TestFunctionQuery, DirectUpdateHandlerTest, TestBadConfig, OverseerCollectionProcessorTest, TestIndexSearcher, HighlighterTest, ShowFileRequestHandlerTest, TestCoreDiscovery, SpellCheckCollatorTest, SuggesterFSTTest, SolrRequestParserTest, TestFoldingMultitermQuery, TestStressLucene, SpatialFilterTest, SuggesterWFSTTest, NoCacheHeaderTest, WordBreakSolrSpellCheckerTest, SolrCoreCheckLockOnStartupTest, TestPseudoReturnFields, TestUpdate, TestWordDelimiterFilterFactory, QueryEqualityTest, DirectUpdateHandlerOptimizeTest, StatelessScriptUpdateProcessorFactoryTest, SortByFunctionTest, TestRemoteStreaming, TestSolrDeletionPolicy1, SolrInfoMBeanTest, XsltUpdateRequestHandlerTest, TestSurroundQueryParser, DisMaxRequestHandlerTest, StandardRequestHandlerTest, DirectSolrSpellCheckerTest, TestReversedWildcardFilterFactory, PrimitiveFieldTypeTest, TestOmitPositions, FileBasedSpellCheckerTest, TermsComponentTest, XmlUpdateRequestHandlerTest, TermVectorComponentTest, TestIndexingPerformance, MoreLikeThisHandlerTest, RequiredFieldsTest, FieldAnalysisRequestHandlerTest, TestArbitraryIndexDir, FastVectorHighlighterTest, IndexSchemaRuntimeFieldTest, LoggingHandlerTest, SolrPluginUtilsTest, TestCollationField, ReturnFieldsTest, JSONWriterTest, TestCSVResponseWriter, MBeansHandlerTest, TestAnalyzedSuggestions, UniqFieldsUpdateProcessorFactoryTest, BinaryUpdateRequestHandlerTest, PingRequestHandlerTest, TestBinaryResponseWriter, TestLFUCache, HighlighterConfigTest, TestPropInject, IndexReaderFactoryTest, UpdateParamsTest, TestSolrIndexConfig, SolrIndexConfigTest, TestStressRecovery, TestMergePolicyConfig, TestSolrDeletionPolicy2, SampleTest, TestDocSet, TestBinaryField, TestSearchPerf, NumericFieldsTest, TestElisionMultitermQuery, OutputWriterTest, DirectSolrConnectionTest, NotRequiredUniqueKeyTest, TestLuceneMatchVersion, SynonymTokenizerTest, EchoParamsTest, TestDFRSimilarityFactory, TestLMDirichletSimilarityFactory, TestLMJelinekMercerSimilarityFactory, TestBM25SimilarityFactory, TestPluginEnable, TimeZoneUtilsTest, ResourceLoaderTest, OpenExchangeRatesOrgProviderTest, TestFastLRUCache, DateMathParserTest, PreAnalyzedFieldTest, SpellingQueryConverterTest, DOMUtilTest, TestLRUCache, TestUtils, TestDocumentBuilder, SliceStateTest, UUIDFieldTest, FileUtilsTest, TestRTGBase, SolrTestCaseJ4Test, AssignTest, DeleteShardTest, ShardSplitTest]
   [junit4] Completed in 77.24s, 1 test, 1 failure <<< FAILURES!

[...truncated 70 lines...]
BUILD FAILED
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:389: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:369: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/build.xml:39: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/build.xml:181: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/solr/common-build.xml:449: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/lucene/common-build.xml:1247: The following error occurred while executing this line:
/Users/jenkins/jenkins-slave/workspace/Lucene-Solr-trunk-MacOSX/lucene/common-build.xml:890: There were test failures: 318 suites, 1342 tests, 2 failures, 39 ignored (5 assumptions)

Total time: 110 minutes 7 seconds
Build step 'Invoke Ant' marked build as failure
Description set: Java: 64bit/jdk1.7.0 -XX:+UseCompressedOops -XX:+UseParallelGC
Archiving artifacts
Recording test results
Email was triggered for: Failure
Sending email for trigger: Failure