You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2015/04/28 17:32:58 UTC

[JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.8.0_60-ea-b12) - Build # 12488 - Failure!

Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/12488/
Java: 32bit/jdk1.8.0_60-ea-b12 -server -XX:+UseConcMarkSweepGC

1 tests failed.
FAILED:  org.apache.solr.cloud.HttpPartitionTest.test

Error Message:
Didn't see all replicas for shard shard1 in c8n_1x2 come up within 30000 ms! ClusterState: {   "collection1":{     "replicationFactor":"1",     "shards":{       "shard1":{         "range":"80000000-ffffffff",         "state":"active",         "replicas":{"core_node2":{             "core":"collection1",             "base_url":"http://127.0.0.1:36422/y_ow",             "node_name":"127.0.0.1:36422_y_ow",             "state":"active",             "leader":"true"}}},       "shard2":{         "range":"0-7fffffff",         "state":"active",         "replicas":{           "core_node1":{             "core":"collection1",             "base_url":"http://127.0.0.1:41627/y_ow",             "node_name":"127.0.0.1:41627_y_ow",             "state":"active",             "leader":"true"},           "core_node3":{             "core":"collection1",             "base_url":"http://127.0.0.1:35144/y_ow",             "node_name":"127.0.0.1:35144_y_ow",             "state":"active"}}}},     "router":{"name":"compositeId"},     "maxShardsPerNode":"1",     "autoAddReplicas":"false",     "autoCreated":"true"},   "control_collection":{     "replicationFactor":"1",     "shards":{"shard1":{         "range":"80000000-7fffffff",         "state":"active",         "replicas":{"core_node1":{             "core":"collection1",             "base_url":"http://127.0.0.1:50216/y_ow",             "node_name":"127.0.0.1:50216_y_ow",             "state":"active",             "leader":"true"}}}},     "router":{"name":"compositeId"},     "maxShardsPerNode":"1",     "autoAddReplicas":"false",     "autoCreated":"true"},   "c8n_1x2":{     "replicationFactor":"2",     "shards":{"shard1":{         "range":"80000000-7fffffff",         "state":"active",         "replicas":{           "core_node1":{             "core":"c8n_1x2_shard1_replica1",             "base_url":"http://127.0.0.1:36422/y_ow",             "node_name":"127.0.0.1:36422_y_ow",             "state":"recovering"},           "core_node2":{             "core":"c8n_1x2_shard1_replica2",             "base_url":"http://127.0.0.1:35144/y_ow",             "node_name":"127.0.0.1:35144_y_ow",             "state":"active",             "leader":"true"}}}},     "router":{"name":"compositeId"},     "maxShardsPerNode":"1",     "autoAddReplicas":"false"}}

Stack Trace:
java.lang.AssertionError: Didn't see all replicas for shard shard1 in c8n_1x2 come up within 30000 ms! ClusterState: {
  "collection1":{
    "replicationFactor":"1",
    "shards":{
      "shard1":{
        "range":"80000000-ffffffff",
        "state":"active",
        "replicas":{"core_node2":{
            "core":"collection1",
            "base_url":"http://127.0.0.1:36422/y_ow",
            "node_name":"127.0.0.1:36422_y_ow",
            "state":"active",
            "leader":"true"}}},
      "shard2":{
        "range":"0-7fffffff",
        "state":"active",
        "replicas":{
          "core_node1":{
            "core":"collection1",
            "base_url":"http://127.0.0.1:41627/y_ow",
            "node_name":"127.0.0.1:41627_y_ow",
            "state":"active",
            "leader":"true"},
          "core_node3":{
            "core":"collection1",
            "base_url":"http://127.0.0.1:35144/y_ow",
            "node_name":"127.0.0.1:35144_y_ow",
            "state":"active"}}}},
    "router":{"name":"compositeId"},
    "maxShardsPerNode":"1",
    "autoAddReplicas":"false",
    "autoCreated":"true"},
  "control_collection":{
    "replicationFactor":"1",
    "shards":{"shard1":{
        "range":"80000000-7fffffff",
        "state":"active",
        "replicas":{"core_node1":{
            "core":"collection1",
            "base_url":"http://127.0.0.1:50216/y_ow",
            "node_name":"127.0.0.1:50216_y_ow",
            "state":"active",
            "leader":"true"}}}},
    "router":{"name":"compositeId"},
    "maxShardsPerNode":"1",
    "autoAddReplicas":"false",
    "autoCreated":"true"},
  "c8n_1x2":{
    "replicationFactor":"2",
    "shards":{"shard1":{
        "range":"80000000-7fffffff",
        "state":"active",
        "replicas":{
          "core_node1":{
            "core":"c8n_1x2_shard1_replica1",
            "base_url":"http://127.0.0.1:36422/y_ow",
            "node_name":"127.0.0.1:36422_y_ow",
            "state":"recovering"},
          "core_node2":{
            "core":"c8n_1x2_shard1_replica2",
            "base_url":"http://127.0.0.1:35144/y_ow",
            "node_name":"127.0.0.1:35144_y_ow",
            "state":"active",
            "leader":"true"}}}},
    "router":{"name":"compositeId"},
    "maxShardsPerNode":"1",
    "autoAddReplicas":"false"}}
	at __randomizedtesting.SeedInfo.seed([375FC4C9173DBE9E:BF0BFB13B9C1D366]:0)
	at org.junit.Assert.fail(Assert.java:93)
	at org.apache.solr.cloud.AbstractFullDistribZkTestBase.ensureAllReplicasAreActive(AbstractFullDistribZkTestBase.java:1920)
	at org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:205)
	at org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:105)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:497)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886)
	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:960)
	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:935)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
	at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781)
	at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
	at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
	at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
	at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
	at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
	at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
	at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
	at java.lang.Thread.run(Thread.java:745)




Build Log:
[...truncated 1709 lines...]
   [junit4] JVM J0: stdout was not empty, see: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/build/core/test/temp/junit4-J0-20150428_145157_996.sysout
   [junit4] >>> JVM J0: stdout (verbatim) ----
   [junit4] Default case invoked for: 
   [junit4]    opcode  = 0, "Node"
   [junit4] Default case invoked for: 
   [junit4]    opcode  = 0, "Node"
   [junit4] Default case invoked for: 
   [junit4]    opcode  = 199, "Phi"
   [junit4] <<< JVM J0: EOF ----

[...truncated 8742 lines...]
   [junit4] Suite: org.apache.solr.cloud.HttpPartitionTest
   [junit4]   2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/init-core-data-001
   [junit4]   2> 960547 T6226 oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system property: /y_ow/
   [junit4]   2> 960549 T6226 oasc.ZkTestServer.run STARTING ZK TEST SERVER
   [junit4]   2> 960549 T6227 oasc.ZkTestServer$2$1.setClientPort client port:0.0.0.0/0.0.0.0:0
   [junit4]   2> 960550 T6227 oasc.ZkTestServer$ZKServerMain.runFromConfig Starting server
   [junit4]   2> 960649 T6226 oasc.ZkTestServer.run start zk server on port:43193
   [junit4]   2> 960663 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml
   [junit4]   2> 960664 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/schema.xml to /configs/conf1/schema.xml
   [junit4]   2> 960665 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
   [junit4]   2> 960666 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
   [junit4]   2> 960667 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
   [junit4]   2> 960668 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
   [junit4]   2> 960669 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml
   [junit4]   2> 960669 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
   [junit4]   2> 960670 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
   [junit4]   2> 960671 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
   [junit4]   2> 960672 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
   [junit4]   2> 960726 T6226 oas.SolrTestCaseJ4.writeCoreProperties Writing core.properties file to /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1
   [junit4]   2> 960728 T6226 oejs.Server.doStart jetty-9.2.10.v20150310
   [junit4]   2> 960729 T6226 oejsh.ContextHandler.doStart Started o.e.j.s.ServletContextHandler@1030194{/y_ow,null,AVAILABLE}
   [junit4]   2> 960735 T6226 oejs.AbstractConnector.doStart Started ServerConnector@1aa997{HTTP/1.1}{127.0.0.1:40508}
   [junit4]   2> 960735 T6226 oejs.Server.doStart Started @961672ms
   [junit4]   2> 960735 T6226 oascse.JettySolrRunner$1.lifeCycleStarted Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/tempDir-001/control/data, hostContext=/y_ow, hostPort=50216, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores}
   [junit4]   2> 960736 T6226 oass.SolrDispatchFilter.init SolrDispatchFilter.init()sun.misc.Launcher$AppClassLoader@e2f2a
   [junit4]   2> 960736 T6226 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/'
   [junit4]   2> 960747 T6226 oasc.SolrXmlConfig.fromFile Loading container configuration from /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/solr.xml
   [junit4]   2> 960758 T6226 oasc.CorePropertiesLocator.<init> Config-defined core root directory: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores
   [junit4]   2> 960759 T6226 oasc.CoreContainer.<init> New CoreContainer 2457809
   [junit4]   2> 960759 T6226 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/]
   [junit4]   2> 960760 T6226 oasc.CoreContainer.load loading shared library: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/lib
   [junit4]   2> 960760 T6226 oasc.SolrResourceLoader.addToClassLoader WARN Can't find (or read) directory to add to classloader: lib (resolved as: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/lib).
   [junit4]   2> 960764 T6226 oashc.HttpShardHandlerFactory.init created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
   [junit4]   2> 960765 T6226 oasu.UpdateShardHandler.<init> Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 960766 T6226 oasl.LogWatcher.createWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
   [junit4]   2> 960766 T6226 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 960766 T6226 oasc.CoreContainer.load Node Name: 127.0.0.1
   [junit4]   2> 960766 T6226 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:43193/solr
   [junit4]   2> 960767 T6226 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 960774 T6226 n:127.0.0.1:50216_y_ow oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50216_y_ow
   [junit4]   2> 960776 T6226 n:127.0.0.1:50216_y_ow oasc.Overseer.close Overseer (id=null) closing
   [junit4]   2> 960777 T6226 n:127.0.0.1:50216_y_ow oasc.OverseerElectionContext.runLeaderProcess I am going to be the leader 127.0.0.1:50216_y_ow
   [junit4]   2> 960777 T6226 n:127.0.0.1:50216_y_ow oasc.Overseer.start Overseer (id=93731867006337027-127.0.0.1:50216_y_ow-n_0000000000) starting
   [junit4]   2> 960780 T6226 n:127.0.0.1:50216_y_ow oasc.OverseerAutoReplicaFailoverThread.<init> Starting OverseerAutoReplicaFailoverThread autoReplicaFailoverWorkLoopDelay=10000 autoReplicaFailoverWaitAfterExpiration=30000 autoReplicaFailoverBadNodeExpiration=60000
   [junit4]   2> 960781 T6256 n:127.0.0.1:50216_y_ow oasc.OverseerCollectionProcessor.run Process current queue of collection creations
   [junit4]   2> 960781 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
   [junit4]   2> 960783 T6226 n:127.0.0.1:50216_y_ow oasc.CorePropertiesLocator.discover Looking for core definitions underneath /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores
   [junit4]   2> 960784 T6226 n:127.0.0.1:50216_y_ow oasc.CoreDescriptor.<init> CORE DESCRIPTOR: {name=collection1, config=solrconfig.xml, transient=false, schema=schema.xml, loadOnStartup=true, instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1, collection=control_collection, absoluteInstDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/, coreNodeName=, dataDir=data/, shard=}
   [junit4]   2> 960784 T6226 n:127.0.0.1:50216_y_ow oasc.CorePropertiesLocator.discoverUnder Found core collection1 in /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/
   [junit4]   2> 960784 T6226 n:127.0.0.1:50216_y_ow oasc.CorePropertiesLocator.discover Found 1 core definitions
   [junit4]   2> 960785 T6258 n:127.0.0.1:50216_y_ow c:control_collection x:collection1 oasc.ZkController.publish publishing core=collection1 state=down collection=control_collection
   [junit4]   2> 960785 T6258 n:127.0.0.1:50216_y_ow c:control_collection x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 960785 T6258 n:127.0.0.1:50216_y_ow oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 960785 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 960786 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50216/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:50216_y_ow",
   [junit4]   2> 	  "numShards":"1",
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "operation":"state"} current state version: 0
   [junit4]   2> 960786 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=1 message={
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50216/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:50216_y_ow",
   [junit4]   2> 	  "numShards":"1",
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "operation":"state"}
   [junit4]   2> 960786 T6255 n:127.0.0.1:50216_y_ow oasco.ClusterStateMutator.createCollection building a new cName: control_collection
   [junit4]   2> 960786 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Assigning new node to shard shard=shard1
   [junit4]   2> 961785 T6258 n:127.0.0.1:50216_y_ow oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 961786 T6258 n:127.0.0.1:50216_y_ow oasc.ZkController.createCollectionZkNode Check for collection zkNode:control_collection
   [junit4]   2> 961786 T6258 n:127.0.0.1:50216_y_ow oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 961787 T6258 n:127.0.0.1:50216_y_ow oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/'
   [junit4]   2> 961796 T6258 n:127.0.0.1:50216_y_ow oasc.Config.<init> loaded config solrconfig.xml with version 0 
   [junit4]   2> 961798 T6258 n:127.0.0.1:50216_y_ow oasc.SolrConfig.refreshRequestParams current version of requestparams : -1
   [junit4]   2> 961801 T6258 n:127.0.0.1:50216_y_ow oasc.SolrConfig.<init> Using Lucene MatchVersion: 6.0.0
   [junit4]   2> 961807 T6258 n:127.0.0.1:50216_y_ow oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 961808 T6258 n:127.0.0.1:50216_y_ow oass.IndexSchema.readSchema Reading Solr Schema from /configs/conf1/schema.xml
   [junit4]   2> 961812 T6258 n:127.0.0.1:50216_y_ow oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 961879 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 961882 T6258 n:127.0.0.1:50216_y_ow oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 961883 T6258 n:127.0.0.1:50216_y_ow oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 961887 T6258 n:127.0.0.1:50216_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 961889 T6258 n:127.0.0.1:50216_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 961890 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 961890 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 961890 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 961890 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 961891 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 961891 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 961891 T6258 n:127.0.0.1:50216_y_ow oasc.CoreContainer.create Creating SolrCore 'collection1' using configuration from collection control_collection
   [junit4]   2> 961891 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 961891 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.<init> [[collection1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/], dataDir=[null]
   [junit4]   2> 961892 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a50e1b
   [junit4]   2> 961892 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/data
   [junit4]   2> 961892 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/data/index/
   [junit4]   2> 961893 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.initIndex WARN [collection1] Solr index directory '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 961893 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/data/index
   [junit4]   2> 961893 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=29, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.18017383033454096]
   [junit4]   2> 961894 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirectoryWrapper(RAMDirectory@880570 lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@1f9bd62),segFN=segments_1,generation=1}
   [junit4]   2> 961894 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 961899 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "nodistrib"
   [junit4]   2> 961899 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "dedupe"
   [junit4]   2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "stored_sig"
   [junit4]   2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-explicit"
   [junit4]   2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 961901 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 961901 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.loadUpdateProcessorChains no updateRequestProcessorChain defined as default, creating implicit default
   [junit4]   2> 961903 T6258 n:127.0.0.1:50216_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 961904 T6258 n:127.0.0.1:50216_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 961904 T6258 n:127.0.0.1:50216_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 961905 T6258 n:127.0.0.1:50216_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 961908 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.RequestHandlers.initHandlersFromConfig Registered paths: /admin/mbeans,standard,/update/csv,/update/json/docs,/admin/luke,/admin/segments,/get,/admin/system,/replication,/admin/properties,/config,/schema,/admin/plugins,/admin/logging,/update/json,/admin/threads,/admin/ping,/update,/admin/file
   [junit4]   2> 961908 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.initStatsCache Using default statsCache cache: org.apache.solr.search.stats.LocalStatsCache
   [junit4]   2> 961909 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.UpdateHandler.<init> Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 961909 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.UpdateLog.init Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10
   [junit4]   2> 961909 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 961910 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 961910 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=46, maxMergeAtOnceExplicit=46, maxMergedSegmentMB=82.8984375, floorSegmentMB=1.099609375, forceMergeDeletesPctAllowed=15.853958222535056, segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
   [junit4]   2> 961910 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirectoryWrapper(RAMDirectory@880570 lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@1f9bd62),segFN=segments_1,generation=1}
   [junit4]   2> 961911 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 961911 T6258 n:127.0.0.1:50216_y_ow x:collection1 oass.SolrIndexSearcher.<init> Opening Searcher@1f42150[collection1] main
   [junit4]   2> 961911 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage.newStorageIO Setting up ZooKeeper-based storage for the RestManager with znodeBase: /configs/conf1
   [junit4]   2> 961912 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.configure Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 961912 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.RestManager.init Initializing RestManager with initArgs: {}
   [junit4]   2> 961912 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage.load Reading _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 961912 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.openInputStream No data found for znode /configs/conf1/_rest_managed.json
   [junit4]   2> 961913 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage.load Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 961913 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.RestManager.init Initializing 0 registered ManagedResources
   [junit4]   2> 961913 T6258 n:127.0.0.1:50216_y_ow x:collection1 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 961913 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.ZkController.getConfDirListeners watch zkdir /configs/conf1
   [junit4]   2> 961913 T6259 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@1f42150[collection1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 961914 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 961914 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50216/y_ow collection:control_collection shard:shard1
   [junit4]   2> 961914 T6226 n:127.0.0.1:50216_y_ow oass.SolrDispatchFilter.init user.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0
   [junit4]   2> 961915 T6226 n:127.0.0.1:50216_y_ow oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 961918 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 961923 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 961923 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 961923 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> 961923 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "operation":"leader",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "collection":"control_collection"} current state version: 1
   [junit4]   2> ASYNC  NEW_CORE C5782 name=collection1 org.apache.solr.core.SolrCore@5387a6 url=http://127.0.0.1:50216/y_ow/collection1 node=127.0.0.1:50216_y_ow C5782_STATE=coll:control_collection core:collection1 props:{core=collection1, base_url=http://127.0.0.1:50216/y_ow, node_name=127.0.0.1:50216_y_ow, state=down}
   [junit4]   2> 961923 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 C5782 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50216/y_ow/collection1/
   [junit4]   2> 961924 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 C5782 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 961924 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 C5782 oasc.SyncStrategy.syncToMe http://127.0.0.1:50216/y_ow/collection1/ has no replicas
   [junit4]   2> 961923 T6226 oasc.ChaosMonkey.monkeyLog monkey: init - expire sessions:false cause connection loss:false
   [junit4]   2> 961924 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50216/y_ow/collection1/ shard1
   [junit4]   2> 961928 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 961928 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "operation":"leader",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50216/y_ow",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "state":"active"} current state version: 1
   [junit4]   2> 961982 T6226 oas.SolrTestCaseJ4.writeCoreProperties Writing core.properties file to /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1
   [junit4]   2> 961983 T6226 oasc.AbstractFullDistribZkTestBase.createJettys create jetty 1 in directory /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001
   [junit4]   2> 961984 T6226 oejs.Server.doStart jetty-9.2.10.v20150310
   [junit4]   2> 961985 T6226 oejsh.ContextHandler.doStart Started o.e.j.s.ServletContextHandler@df06d0{/y_ow,null,AVAILABLE}
   [junit4]   2> 961985 T6226 oejs.AbstractConnector.doStart Started ServerConnector@1cebdab{HTTP/1.1}{127.0.0.1:35724}
   [junit4]   2> 961985 T6226 oejs.Server.doStart Started @962922ms
   [junit4]   2> 961986 T6226 oascse.JettySolrRunner$1.lifeCycleStarted Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/tempDir-001/jetty1, solrconfig=solrconfig.xml, hostContext=/y_ow, hostPort=41627, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores}
   [junit4]   2> 961986 T6226 oass.SolrDispatchFilter.init SolrDispatchFilter.init()sun.misc.Launcher$AppClassLoader@e2f2a
   [junit4]   2> 961986 T6226 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/'
   [junit4]   2> 961995 T6226 oasc.SolrXmlConfig.fromFile Loading container configuration from /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/solr.xml
   [junit4]   2> 961999 T6226 oasc.CorePropertiesLocator.<init> Config-defined core root directory: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores
   [junit4]   2> 962000 T6226 oasc.CoreContainer.<init> New CoreContainer 5753138
   [junit4]   2> 962000 T6226 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/]
   [junit4]   2> 962000 T6226 oasc.CoreContainer.load loading shared library: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/lib
   [junit4]   2> 962000 T6226 oasc.SolrResourceLoader.addToClassLoader WARN Can't find (or read) directory to add to classloader: lib (resolved as: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/lib).
   [junit4]   2> 962004 T6226 oashc.HttpShardHandlerFactory.init created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
   [junit4]   2> 962005 T6226 oasu.UpdateShardHandler.<init> Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 962006 T6226 oasl.LogWatcher.createWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
   [junit4]   2> 962006 T6226 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 962006 T6226 oasc.CoreContainer.load Node Name: 127.0.0.1
   [junit4]   2> 962006 T6226 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:43193/solr
   [junit4]   2> 962007 T6226 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 962078 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.register We are http://127.0.0.1:50216/y_ow/collection1/ and leader is http://127.0.0.1:50216/y_ow/collection1/
   [junit4]   2> 962079 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:50216/y_ow
   [junit4]   2> 962079 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 962079 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.publish publishing core=collection1 state=active collection=control_collection
   [junit4]   2> 962079 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 962080 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 962080 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50216/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:50216_y_ow",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "operation":"state"} current state version: 2
   [junit4]   2> 962081 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=2 message={
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:50216/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:50216_y_ow",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "collection":"control_collection",
   [junit4]   2> 	  "operation":"state"}
   [junit4]   2> 963013 T6226 n:127.0.0.1:41627_y_ow oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:41627_y_ow
   [junit4]   2> 963015 T6226 n:127.0.0.1:41627_y_ow oasc.Overseer.close Overseer (id=null) closing
   [junit4]   2> 963016 T6226 n:127.0.0.1:41627_y_ow oasc.CorePropertiesLocator.discover Looking for core definitions underneath /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores
   [junit4]   2> 963017 T6226 n:127.0.0.1:41627_y_ow oasc.CoreDescriptor.<init> CORE DESCRIPTOR: {name=collection1, config=solrconfig.xml, transient=false, schema=schema.xml, loadOnStartup=true, instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1, collection=collection1, absoluteInstDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/, coreNodeName=, dataDir=data/, shard=}
   [junit4]   2> 963017 T6226 n:127.0.0.1:41627_y_ow oasc.CorePropertiesLocator.discoverUnder Found core collection1 in /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/
   [junit4]   2> 963018 T6226 n:127.0.0.1:41627_y_ow oasc.CorePropertiesLocator.discover Found 1 core definitions
   [junit4]   2> 963019 T6284 n:127.0.0.1:41627_y_ow c:collection1 x:collection1 oasc.ZkController.publish publishing core=collection1 state=down collection=collection1
   [junit4]   2> 963019 T6284 n:127.0.0.1:41627_y_ow c:collection1 x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 963019 T6284 n:127.0.0.1:41627_y_ow oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 963019 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 963020 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:41627/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:41627_y_ow",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "operation":"state"} current state version: 3
   [junit4]   2> 963020 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=2 message={
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:41627/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:41627_y_ow",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "operation":"state"}
   [junit4]   2> 963021 T6255 n:127.0.0.1:50216_y_ow oasco.ClusterStateMutator.createCollection building a new cName: collection1
   [junit4]   2> 963021 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Assigning new node to shard shard=shard2
   [junit4]   2> 964020 T6284 n:127.0.0.1:41627_y_ow oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 964020 T6284 n:127.0.0.1:41627_y_ow oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 964020 T6284 n:127.0.0.1:41627_y_ow oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 964021 T6284 n:127.0.0.1:41627_y_ow oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/'
   [junit4]   2> 964033 T6284 n:127.0.0.1:41627_y_ow oasc.Config.<init> loaded config solrconfig.xml with version 0 
   [junit4]   2> 964038 T6284 n:127.0.0.1:41627_y_ow oasc.SolrConfig.refreshRequestParams current version of requestparams : -1
   [junit4]   2> 964041 T6284 n:127.0.0.1:41627_y_ow oasc.SolrConfig.<init> Using Lucene MatchVersion: 6.0.0
   [junit4]   2> 964068 T6284 n:127.0.0.1:41627_y_ow oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 964072 T6284 n:127.0.0.1:41627_y_ow oass.IndexSchema.readSchema Reading Solr Schema from /configs/conf1/schema.xml
   [junit4]   2> 964076 T6284 n:127.0.0.1:41627_y_ow oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 964149 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 964154 T6284 n:127.0.0.1:41627_y_ow oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 964155 T6284 n:127.0.0.1:41627_y_ow oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 964159 T6284 n:127.0.0.1:41627_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 964160 T6284 n:127.0.0.1:41627_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 964161 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 964162 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 964162 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 964162 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 964164 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 964165 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 964165 T6284 n:127.0.0.1:41627_y_ow oasc.CoreContainer.create Creating SolrCore 'collection1' using configuration from collection collection1
   [junit4]   2> 964165 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 964165 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.<init> [[collection1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/], dataDir=[null]
   [junit4]   2> 964165 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a50e1b
   [junit4]   2> 964166 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/data
   [junit4]   2> 964166 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/data/index/
   [junit4]   2> 964166 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.initIndex WARN [collection1] Solr index directory '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 964166 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/data/index
   [junit4]   2> 964167 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=29, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.18017383033454096]
   [junit4]   2> 964167 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirectoryWrapper(RAMDirectory@b4ca8f lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@4bb787),segFN=segments_1,generation=1}
   [junit4]   2> 964167 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 964169 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "nodistrib"
   [junit4]   2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "dedupe"
   [junit4]   2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "stored_sig"
   [junit4]   2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-explicit"
   [junit4]   2> 964171 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 964171 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 964171 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.loadUpdateProcessorChains no updateRequestProcessorChain defined as default, creating implicit default
   [junit4]   2> 964172 T6284 n:127.0.0.1:41627_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 964173 T6284 n:127.0.0.1:41627_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 964174 T6284 n:127.0.0.1:41627_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 964174 T6284 n:127.0.0.1:41627_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 964178 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.RequestHandlers.initHandlersFromConfig Registered paths: /admin/mbeans,standard,/update/csv,/update/json/docs,/admin/luke,/admin/segments,/get,/admin/system,/replication,/admin/properties,/config,/schema,/admin/plugins,/admin/logging,/update/json,/admin/threads,/admin/ping,/update,/admin/file
   [junit4]   2> 964179 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.initStatsCache Using default statsCache cache: org.apache.solr.search.stats.LocalStatsCache
   [junit4]   2> 964179 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.UpdateHandler.<init> Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 964179 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.UpdateLog.init Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10
   [junit4]   2> 964180 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 964180 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 964181 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=46, maxMergeAtOnceExplicit=46, maxMergedSegmentMB=82.8984375, floorSegmentMB=1.099609375, forceMergeDeletesPctAllowed=15.853958222535056, segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
   [junit4]   2> 964181 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirectoryWrapper(RAMDirectory@b4ca8f lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@4bb787),segFN=segments_1,generation=1}
   [junit4]   2> 964181 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 964181 T6284 n:127.0.0.1:41627_y_ow x:collection1 oass.SolrIndexSearcher.<init> Opening Searcher@1f712d6[collection1] main
   [junit4]   2> 964182 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage.newStorageIO Setting up ZooKeeper-based storage for the RestManager with znodeBase: /configs/conf1
   [junit4]   2> 964182 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.configure Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 964182 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.RestManager.init Initializing RestManager with initArgs: {}
   [junit4]   2> 964182 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage.load Reading _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 964183 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.openInputStream No data found for znode /configs/conf1/_rest_managed.json
   [junit4]   2> 964183 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage.load Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 964183 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.RestManager.init Initializing 0 registered ManagedResources
   [junit4]   2> 964183 T6284 n:127.0.0.1:41627_y_ow x:collection1 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 964184 T6285 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@1f712d6[collection1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 964184 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.ZkController.getConfDirListeners watch zkdir /configs/conf1
   [junit4]   2> 964184 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 964200 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:41627/y_ow collection:collection1 shard:shard2
   [junit4]   2> 964200 T6226 n:127.0.0.1:41627_y_ow oass.SolrDispatchFilter.init user.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0
   [junit4]   2> 964200 T6226 n:127.0.0.1:41627_y_ow oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 964203 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
   [junit4]   2> 964204 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 964204 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> ASYNC  NEW_CORE C5783 name=collection1 org.apache.solr.core.SolrCore@6da5c2 url=http://127.0.0.1:41627/y_ow/collection1 node=127.0.0.1:41627_y_ow C5783_STATE=coll:collection1 core:collection1 props:{core=collection1, base_url=http://127.0.0.1:41627/y_ow, node_name=127.0.0.1:41627_y_ow, state=down}
   [junit4]   2> 964204 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 C5783 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:41627/y_ow/collection1/
   [junit4]   2> 964204 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 C5783 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 964205 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 C5783 oasc.SyncStrategy.syncToMe http://127.0.0.1:41627/y_ow/collection1/ has no replicas
   [junit4]   2> 964205 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:41627/y_ow/collection1/ shard2
   [junit4]   2> 964206 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 964207 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 2, message = {
   [junit4]   2> 	  "operation":"leader",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "collection":"collection1"} current state version: 4
   [junit4]   2> 964211 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "operation":"leader",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:41627/y_ow",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "state":"active"} current state version: 4
   [junit4]   2> 964212 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 964256 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.register We are http://127.0.0.1:41627/y_ow/collection1/ and leader is http://127.0.0.1:41627/y_ow/collection1/
   [junit4]   2> 964257 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:41627/y_ow
   [junit4]   2> 964257 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
   [junit4]   2> 964257 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.publish publishing core=collection1 state=active collection=collection1
   [junit4]   2> 964257 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 964258 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 964263 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:41627/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:41627_y_ow",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "operation":"state"} current state version: 5
   [junit4]   2> 964265 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=2 message={
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "core_node_name":"core_node1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:41627/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:41627_y_ow",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "state":"active",
   [junit4]   2> 	  "shard":"shard2",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "operation":"state"}
   [junit4]   2> 964276 T6226 oas.SolrTestCaseJ4.writeCoreProperties Writing core.properties file to /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1
   [junit4]   2> 964277 T6226 oasc.AbstractFullDistribZkTestBase.createJettys create jetty 2 in directory /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001
   [junit4]   2> 964278 T6226 oejs.Server.doStart jetty-9.2.10.v20150310
   [junit4]   2> 964279 T6226 oejsh.ContextHandler.doStart Started o.e.j.s.ServletContextHandler@bb9abb{/y_ow,null,AVAILABLE}
   [junit4]   2> 964279 T6226 oejs.AbstractConnector.doStart Started ServerConnector@16435ad{HTTP/1.1}{127.0.0.1:40120}
   [junit4]   2> 964280 T6226 oejs.Server.doStart Started @965216ms
   [junit4]   2> 964280 T6226 oascse.JettySolrRunner$1.lifeCycleStarted Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/tempDir-001/jetty2, solrconfig=solrconfig.xml, hostContext=/y_ow, hostPort=36422, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores}
   [junit4]   2> 964280 T6226 oass.SolrDispatchFilter.init SolrDispatchFilter.init()sun.misc.Launcher$AppClassLoader@e2f2a
   [junit4]   2> 964281 T6226 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/'
   [junit4]   2> 964291 T6226 oasc.SolrXmlConfig.fromFile Loading container configuration from /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/solr.xml
   [junit4]   2> 964300 T6226 oasc.CorePropertiesLocator.<init> Config-defined core root directory: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores
   [junit4]   2> 964300 T6226 oasc.CoreContainer.<init> New CoreContainer 12242390
   [junit4]   2> 964301 T6226 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/]
   [junit4]   2> 964301 T6226 oasc.CoreContainer.load loading shared library: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/lib
   [junit4]   2> 964301 T6226 oasc.SolrResourceLoader.addToClassLoader WARN Can't find (or read) directory to add to classloader: lib (resolved as: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/lib).
   [junit4]   2> 964306 T6226 oashc.HttpShardHandlerFactory.init created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
   [junit4]   2> 964308 T6226 oasu.UpdateShardHandler.<init> Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 964309 T6226 oasl.LogWatcher.createWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
   [junit4]   2> 964309 T6226 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
   [junit4]   2> 964310 T6226 oasc.CoreContainer.load Node Name: 127.0.0.1
   [junit4]   2> 964310 T6226 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:43193/solr
   [junit4]   2> 964310 T6226 oasc.ZkController.checkChrootPath zkHost includes chroot
   [junit4]   2> 965325 T6226 n:127.0.0.1:36422_y_ow oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:36422_y_ow
   [junit4]   2> 965327 T6226 n:127.0.0.1:36422_y_ow oasc.Overseer.close Overseer (id=null) closing
   [junit4]   2> 965328 T6226 n:127.0.0.1:36422_y_ow oasc.CorePropertiesLocator.discover Looking for core definitions underneath /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores
   [junit4]   2> 965329 T6226 n:127.0.0.1:36422_y_ow oasc.CoreDescriptor.<init> CORE DESCRIPTOR: {name=collection1, config=solrconfig.xml, transient=false, schema=schema.xml, loadOnStartup=true, instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1, collection=collection1, absoluteInstDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/, coreNodeName=, dataDir=data/, shard=}
   [junit4]   2> 965329 T6226 n:127.0.0.1:36422_y_ow oasc.CorePropertiesLocator.discoverUnder Found core collection1 in /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/
   [junit4]   2> 965330 T6226 n:127.0.0.1:36422_y_ow oasc.CorePropertiesLocator.discover Found 1 core definitions
   [junit4]   2> 965331 T6307 n:127.0.0.1:36422_y_ow c:collection1 x:collection1 oasc.ZkController.publish publishing core=collection1 state=down collection=collection1
   [junit4]   2> 965332 T6307 n:127.0.0.1:36422_y_ow c:collection1 x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
   [junit4]   2> 965333 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 965333 T6307 n:127.0.0.1:36422_y_ow oasc.ZkController.waitForCoreNodeName look for our core node name
   [junit4]   2> 965333 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:36422/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:36422_y_ow",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "operation":"state"} current state version: 6
   [junit4]   2> 965334 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=2 message={
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "roles":null,
   [junit4]   2> 	  "base_url":"http://127.0.0.1:36422/y_ow",
   [junit4]   2> 	  "node_name":"127.0.0.1:36422_y_ow",
   [junit4]   2> 	  "numShards":"2",
   [junit4]   2> 	  "state":"down",
   [junit4]   2> 	  "shard":null,
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "operation":"state"}
   [junit4]   2> 965334 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Collection already exists with numShards=2
   [junit4]   2> 965334 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Assigning new node to shard shard=shard1
   [junit4]   2> 966333 T6307 n:127.0.0.1:36422_y_ow oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
   [junit4]   2> 966333 T6307 n:127.0.0.1:36422_y_ow oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
   [junit4]   2> 966334 T6307 n:127.0.0.1:36422_y_ow oasc.ZkController.createCollectionZkNode Collection zkNode exists
   [junit4]   2> 966334 T6307 n:127.0.0.1:36422_y_ow oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/'
   [junit4]   2> 966342 T6307 n:127.0.0.1:36422_y_ow oasc.Config.<init> loaded config solrconfig.xml with version 0 
   [junit4]   2> 966345 T6307 n:127.0.0.1:36422_y_ow oasc.SolrConfig.refreshRequestParams current version of requestparams : -1
   [junit4]   2> 966348 T6307 n:127.0.0.1:36422_y_ow oasc.SolrConfig.<init> Using Lucene MatchVersion: 6.0.0
   [junit4]   2> 966359 T6307 n:127.0.0.1:36422_y_ow oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
   [junit4]   2> 966360 T6307 n:127.0.0.1:36422_y_ow oass.IndexSchema.readSchema Reading Solr Schema from /configs/conf1/schema.xml
   [junit4]   2> 966363 T6307 n:127.0.0.1:36422_y_ow oass.IndexSchema.readSchema [collection1] Schema name=test
   [junit4]   2> 966462 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
   [junit4]   2> 966467 T6307 n:127.0.0.1:36422_y_ow oass.IndexSchema.readSchema default search field in schema is text
   [junit4]   2> 966468 T6307 n:127.0.0.1:36422_y_ow oass.IndexSchema.readSchema unique key field: id
   [junit4]   2> 966472 T6307 n:127.0.0.1:36422_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 966474 T6307 n:127.0.0.1:36422_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
   [junit4]   2> 966475 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 966475 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 966476 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 966476 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
   [junit4]   2> 966476 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
   [junit4]   2> 966476 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
   [junit4]   2> 966476 T6307 n:127.0.0.1:36422_y_ow oasc.CoreContainer.create Creating SolrCore 'collection1' using configuration from collection collection1
   [junit4]   2> 966477 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
   [junit4]   2> 966477 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.<init> [[collection1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/], dataDir=[null]
   [junit4]   2> 966477 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a50e1b
   [junit4]   2> 966477 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/data
   [junit4]   2> 966478 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/data/index/
   [junit4]   2> 966478 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.initIndex WARN [collection1] Solr index directory '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/data/index' doesn't exist. Creating new index...
   [junit4]   2> 966478 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/data/index
   [junit4]   2> 966478 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=29, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.18017383033454096]
   [junit4]   2> 966479 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirectoryWrapper(RAMDirectory@c5953b lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@e6eed5),segFN=segments_1,generation=1}
   [junit4]   2> 966479 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 966481 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "nodistrib"
   [junit4]   2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "dedupe"
   [junit4]   2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
   [junit4]   2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "stored_sig"
   [junit4]   2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
   [junit4]   2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-explicit"
   [junit4]   2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 966483 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
   [junit4]   2> 966483 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.loadUpdateProcessorChains no updateRequestProcessorChain defined as default, creating implicit default
   [junit4]   2> 966484 T6307 n:127.0.0.1:36422_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 966485 T6307 n:127.0.0.1:36422_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 966485 T6307 n:127.0.0.1:36422_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 966486 T6307 n:127.0.0.1:36422_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
   [junit4]   2> 966489 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.RequestHandlers.initHandlersFromConfig Registered paths: /admin/mbeans,standard,/update/csv,/update/json/docs,/admin/luke,/admin/segments,/get,/admin/system,/replication,/admin/properties,/config,/schema,/admin/plugins,/admin/logging,/update/json,/admin/threads,/admin/ping,/update,/admin/file
   [junit4]   2> 966489 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.initStatsCache Using default statsCache cache: org.apache.solr.search.stats.LocalStatsCache
   [junit4]   2> 966490 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.UpdateHandler.<init> Using UpdateLog implementation: org.apache.solr.update.UpdateLog
   [junit4]   2> 966490 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.UpdateLog.init Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10
   [junit4]   2> 966490 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.CommitTracker.<init> Hard AutoCommit: disabled
   [junit4]   2> 966491 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.CommitTracker.<init> Soft AutoCommit: disabled
   [junit4]   2> 966492 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=46, maxMergeAtOnceExplicit=46, maxMergedSegmentMB=82.8984375, floorSegmentMB=1.099609375, forceMergeDeletesPctAllowed=15.853958222535056, segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
   [junit4]   2> 966492 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
   [junit4]   2> 		commit{dir=MockDirectoryWrapper(RAMDirectory@c5953b lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@e6eed5),segFN=segments_1,generation=1}
   [junit4]   2> 966492 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
   [junit4]   2> 966492 T6307 n:127.0.0.1:36422_y_ow x:collection1 oass.SolrIndexSearcher.<init> Opening Searcher@1feb4da[collection1] main
   [junit4]   2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage.newStorageIO Setting up ZooKeeper-based storage for the RestManager with znodeBase: /configs/conf1
   [junit4]   2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.configure Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
   [junit4]   2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.RestManager.init Initializing RestManager with initArgs: {}
   [junit4]   2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage.load Reading _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.openInputStream No data found for znode /configs/conf1/_rest_managed.json
   [junit4]   2> 966494 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage.load Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
   [junit4]   2> 966494 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.RestManager.init Initializing 0 registered ManagedResources
   [junit4]   2> 966494 T6307 n:127.0.0.1:36422_y_ow x:collection1 oash.ReplicationHandler.inform Commits will be reserved for  10000
   [junit4]   2> 966494 T6308 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@1feb4da[collection1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
   [junit4]   2> 966495 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.ZkController.getConfDirListeners watch zkdir /configs/conf1
   [junit4]   2> 966495 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.CoreContainer.registerCore registering core: collection1
   [junit4]   2> 966495 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:36422/y_ow collection:collection1 shard:shard1
   [junit4]   2> 966495 T6226 n:127.0.0.1:36422_y_ow oass.SolrDispatchFilter.init user.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0
   [junit4]   2> 966496 T6226 n:127.0.0.1:36422_y_ow oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
   [junit4]   2> 966497 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
   [junit4]   2> 966498 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 966498 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
   [junit4]   2> 966498 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
   [junit4]   2> ASYNC  NEW_CORE C5784 name=collection1 org.apache.solr.core.SolrCore@d15410 url=http://127.0.0.1:36422/y_ow/collection1 node=127.0.0.1:36422_y_ow C5784_STATE=coll:collection1 core:collection1 props:{core=collection1, base_url=http://127.0.0.1:36422/y_ow, node_name=127.0.0.1:36422_y_ow, state=down}
   [junit4]   2> 966498 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 C5784 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:36422/y_ow/collection1/
   [junit4]   2> 966498 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "operation":"leader",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "collection":"collection1"} current state version: 7
   [junit4]   2> 966498 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 C5784 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
   [junit4]   2> 966499 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 C5784 oasc.SyncStrategy.syncToMe http://127.0.0.1:36422/y_ow/collection1/ has no replicas
   [junit4]   2> 966499 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:36422/y_ow/collection1/ shard1
   [junit4]   2> 966500 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
   [junit4]   2> 966500 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
   [junit4]   2> 	  "operation":"leader",
   [junit4]   2> 	  "shard":"shard1",
   [junit4]   2> 	  "collection":"collection1",
   [junit4]   2> 	  "base_url":"http://127.0.0.1:36422/y_ow",
   [junit4]   2> 	  "core":"collection1",
   [junit4]   2> 	  "state":"active"} current state version: 7
   [junit4]   2> 966556 T6226 oas.SolrTestCaseJ4.writeCoreProperties Writing core.properties file to /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/cores/collection1
   [junit4]   2> 966557 T6226 oasc.AbstractFullDistribZkTestBase.createJettys create jetty 3 in directory /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001
   [junit4]   2> 966558 T6226 oejs.Server.doStart jetty-9.2.10.v20150310
   [junit4]   2> 966559 T6226 oejsh.ContextHandler.doStart Started o.e.j.s.ServletContextHandler@78346c{/y_ow,null,AVAILABLE}
   [junit4]   2> 966560 T6226 oejs.AbstractConnector.doStart Started ServerConnector@771834{HTTP/1.1}{127.0.0.1:37028}
   [junit4]   2> 966560 T6226 oejs.Server.doStart Started @967496ms
   [junit4]   2> 966560 T6226 oascse.JettySolrRunner$1.lifeCycleStarted Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/tempDir-001/jetty3, solrconfig=solrconfig.xml, hostContext=/y_ow, hostPort=35144, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/cores}
   [junit4]   2> 966560 T6226 oass.SolrDispatchFilter.init SolrDispatchFilter.init()sun.misc.Launcher$AppClassLoader@e2f2a
   [junit4]   2> 966561 T6226 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/'
   [junit4]   2> 966570 T6226 oasc.SolrXmlConfig.fromFile Loading container configuration from /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/solr.xml
   [junit4]   2> 966573 T6226 oasc.CorePropertiesLocator.<init> Config-defined core root directory: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/cores
   [junit4]   2> 966573 T6226 oasc.CoreContainer.<init> New CoreContainer 31207983
   [junit4]   2> 966574 T6226 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/]
   [junit4]   2> 966574 T6226 oasc.CoreContainer.load loading shared library: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/lib
   [junit4]   2> 966574 T6226 oasc.SolrResourceLoader.addToClassLoader WARN Can't find (or read) directory to add to classloader: lib (resolved as: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/lib).
   [junit4]   2> 966579 T6226 oashc.HttpShardHandlerFactory.init created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
   [junit4]   2> 966580 T6226 oasu.UpdateShardHandler.<init> Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
   [junit4]   2> 966580 T6226 oasl.LogWatcher.createWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
   [junit4]   2> 966580 T6226 oasl.LogWatcher.newRegisteredLogWatcher Registering Log L

[...truncated too long message...]

  [junit4]   2> 		5	/solr/aliases.json
   [junit4]   2> 		5	/solr/clusterstate.json
   [junit4]   2> 		4	/solr/configs/conf1
   [junit4]   2> 	
   [junit4]   2> 	Maximum concurrent children watches above limit:
   [junit4]   2> 	
   [junit4]   2> 		5	/solr/live_nodes
   [junit4]   2> 		4	/solr/overseer/queue
   [junit4]   2> 		4	/solr/overseer/collection-queue-work
   [junit4]   2> 	
   [junit4]   2> 1034796 T6226 c:control_collection s:shard1 x:collection1 oasc.SocketProxy.close WARN Closing 10 connections to: http://127.0.0.1:41627/y_ow, target: http://127.0.0.1:35724/y_ow
   [junit4]   2> 1034797 T6226 c:control_collection s:shard1 x:collection1 oasc.SocketProxy.close WARN Closing 4 connections to: http://127.0.0.1:50216/y_ow, target: http://127.0.0.1:40508/y_ow
   [junit4]   2> 1034797 T6226 c:control_collection s:shard1 x:collection1 oasc.SocketProxy.close WARN Closing 6 connections to: http://127.0.0.1:35144/y_ow, target: http://127.0.0.1:37028/y_ow
   [junit4]   2> 1034797 T6226 c:control_collection s:shard1 x:collection1 oasc.SocketProxy.close WARN Closing 1 connections to: http://127.0.0.1:36422/y_ow, target: http://127.0.0.1:40120/y_ow
   [junit4]   2> NOTE: reproduce with: ant test  -Dtestcase=HttpPartitionTest -Dtests.method=test -Dtests.seed=375FC4C9173DBE9E -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ar_SD -Dtests.timezone=Atlantic/Jan_Mayen -Dtests.asserts=true -Dtests.file.encoding=UTF-8
   [junit4] FAILURE 74.3s J0 | HttpPartitionTest.test <<<
   [junit4]    > Throwable #1: java.lang.AssertionError: Didn't see all replicas for shard shard1 in c8n_1x2 come up within 30000 ms! ClusterState: {
   [junit4]    >   "collection1":{
   [junit4]    >     "replicationFactor":"1",
   [junit4]    >     "shards":{
   [junit4]    >       "shard1":{
   [junit4]    >         "range":"80000000-ffffffff",
   [junit4]    >         "state":"active",
   [junit4]    >         "replicas":{"core_node2":{
   [junit4]    >             "core":"collection1",
   [junit4]    >             "base_url":"http://127.0.0.1:36422/y_ow",
   [junit4]    >             "node_name":"127.0.0.1:36422_y_ow",
   [junit4]    >             "state":"active",
   [junit4]    >             "leader":"true"}}},
   [junit4]    >       "shard2":{
   [junit4]    >         "range":"0-7fffffff",
   [junit4]    >         "state":"active",
   [junit4]    >         "replicas":{
   [junit4]    >           "core_node1":{
   [junit4]    >             "core":"collection1",
   [junit4]    >             "base_url":"http://127.0.0.1:41627/y_ow",
   [junit4]    >             "node_name":"127.0.0.1:41627_y_ow",
   [junit4]    >             "state":"active",
   [junit4]    >             "leader":"true"},
   [junit4]    >           "core_node3":{
   [junit4]    >             "core":"collection1",
   [junit4]    >             "base_url":"http://127.0.0.1:35144/y_ow",
   [junit4]    >             "node_name":"127.0.0.1:35144_y_ow",
   [junit4]    >             "state":"active"}}}},
   [junit4]    >     "router":{"name":"compositeId"},
   [junit4]    >     "maxShardsPerNode":"1",
   [junit4]    >     "autoAddReplicas":"false",
   [junit4]    >     "autoCreated":"true"},
   [junit4]    >   "control_collection":{
   [junit4]    >     "replicationFactor":"1",
   [junit4]    >     "shards":{"shard1":{
   [junit4]    >         "range":"80000000-7fffffff",
   [junit4]    >         "state":"active",
   [junit4]    >         "replicas":{"core_node1":{
   [junit4]    >             "core":"collection1",
   [junit4]    >             "base_url":"http://127.0.0.1:50216/y_ow",
   [junit4]    >             "node_name":"127.0.0.1:50216_y_ow",
   [junit4]    >             "state":"active",
   [junit4]    >             "leader":"true"}}}},
   [junit4]    >     "router":{"name":"compositeId"},
   [junit4]    >     "maxShardsPerNode":"1",
   [junit4]    >     "autoAddReplicas":"false",
   [junit4]    >     "autoCreated":"true"},
   [junit4]    >   "c8n_1x2":{
   [junit4]    >     "replicationFactor":"2",
   [junit4]    >     "shards":{"shard1":{
   [junit4]    >         "range":"80000000-7fffffff",
   [junit4]    >         "state":"active",
   [junit4]    >         "replicas":{
   [junit4]    >           "core_node1":{
   [junit4]    >             "core":"c8n_1x2_shard1_replica1",
   [junit4]    >             "base_url":"http://127.0.0.1:36422/y_ow",
   [junit4]    >             "node_name":"127.0.0.1:36422_y_ow",
   [junit4]    >             "state":"recovering"},
   [junit4]    >           "core_node2":{
   [junit4]    >             "core":"c8n_1x2_shard1_replica2",
   [junit4]    >             "base_url":"http://127.0.0.1:35144/y_ow",
   [junit4]    >             "node_name":"127.0.0.1:35144_y_ow",
   [junit4]    >             "state":"active",
   [junit4]    >             "leader":"true"}}}},
   [junit4]    >     "router":{"name":"compositeId"},
   [junit4]    >     "maxShardsPerNode":"1",
   [junit4]    >     "autoAddReplicas":"false"}}
   [junit4]    > 	at __randomizedtesting.SeedInfo.seed([375FC4C9173DBE9E:BF0BFB13B9C1D366]:0)
   [junit4]    > 	at org.apache.solr.cloud.AbstractFullDistribZkTestBase.ensureAllReplicasAreActive(AbstractFullDistribZkTestBase.java:1920)
   [junit4]    > 	at org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:205)
   [junit4]    > 	at org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:105)
   [junit4]    > 	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:960)
   [junit4]    > 	at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:935)
   [junit4]    > 	at java.lang.Thread.run(Thread.java:745)
   [junit4]   2> 1034804 T6226 c:control_collection s:shard1 x:collection1 oas.SolrTestCaseJ4.deleteCore ###deleteCore
   [junit4]   2> NOTE: leaving temporary files on disk at: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001
   [junit4]   2> 74259 T6225 ccr.ThreadLeakControl.checkThreadLeaks WARNING Will linger awaiting termination of 1 leaked thread(s).
   [junit4]   2> NOTE: test params are: codec=Asserting(Lucene50): {range_facet_l_dv=Lucene50(blocksize=128), _version_=PostingsFormat(name=Memory doPackFST= true), multiDefault=PostingsFormat(name=MockRandom), a_t=PostingsFormat(name=MockRandom), intDefault=PostingsFormat(name=Memory doPackFST= true), id=Lucene50(blocksize=128), range_facet_i_dv=PostingsFormat(name=MockRandom), text=PostingsFormat(name=Asserting), range_facet_l=PostingsFormat(name=MockRandom), timestamp=PostingsFormat(name=MockRandom)}, docValues:{range_facet_l_dv=DocValuesFormat(name=Asserting), range_facet_i_dv=DocValuesFormat(name=Lucene50), timestamp=DocValuesFormat(name=Lucene50)}, sim=RandomSimilarityProvider(queryNorm=false,coord=yes): {}, locale=ar_SD, timezone=Atlantic/Jan_Mayen
   [junit4]   2> NOTE: Linux 3.13.0-49-generic i386/Oracle Corporation 1.8.0_60-ea (32-bit)/cpus=12,threads=1,free=83234120,total=205910016
   [junit4]   2> NOTE: All tests run in this JVM: [ZkNodePropsTest, SuggesterFSTTest, HighlighterConfigTest, SimpleCollectionCreateDeleteTest, SchemaVersionSpecificBehaviorTest, PeerSyncTest, ExternalFileFieldSortTest, TestStressLucene, TestJmxIntegration, MBeansHandlerTest, DebugComponentTest, TestJoin, TestSolrConfigHandler, QueryElevationComponentTest, TestReversedWildcardFilterFactory, TestConfigOverlay, TestRandomDVFaceting, TestJettySolrRunner, TestCodecSupport, SyncSliceTest, TestDynamicFieldCollectionResource, TestQuerySenderListener, TestRealTimeGet, ChaosMonkeyNothingIsSafeTest, DefaultValueUpdateProcessorTest, TestRangeQuery, DateFieldTest, DistributedFacetPivotSmallTest, SolrCloudExampleTest, TimeZoneUtilsTest, SpatialRPTFieldTypeTest, TestIBSimilarityFactory, TestMiniSolrCloudCluster, TestInfoStreamLogging, TestStandardQParsers, SharedFSAutoReplicaFailoverTest, TestSolr4Spatial, TermVectorComponentTest, LeaderElectionIntegrationTest, HdfsSyncSliceTest, TestDistributedMissingSort, TestRemoteStreaming, ActionThrottleTest, TestConfig, TestMiniSolrCloudClusterSSL, BadCopyFieldTest, TestBlobHandler, BasicDistributedZkTest, TestBadConfig, TestDynamicLoading, TestTrackingShardHandlerFactory, MultiTermTest, TestRandomFaceting, TestSerializedLuceneMatchVersion, TestMergePolicyConfig, TestBulkSchemaConcurrent, ShardSplitTest, TestRecoveryHdfs, TestSolrConfigHandlerCloud, ClusterStateTest, SimpleMLTQParserTest, QueryResultKeyTest, TestReplicationHandlerBackup, TestLRUStatsCache, HttpPartitionTest]
   [junit4] Completed [364/485] on J0 in 75.02s, 1 test, 1 failure <<< FAILURES!

[...truncated 370 lines...]
BUILD FAILED
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:526: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:474: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:61: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/extra-targets.xml:39: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build.xml:229: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/common-build.xml:512: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/common-build.xml:1415: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/common-build.xml:973: There were test failures: 485 suites, 1947 tests, 1 failure, 57 ignored (25 assumptions)

Total time: 41 minutes 13 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any