You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@lucene.apache.org by Policeman Jenkins Server <je...@thetaphi.de> on 2015/04/28 17:32:58 UTC
[JENKINS] Lucene-Solr-trunk-Linux (32bit/jdk1.8.0_60-ea-b12) -
Build # 12488 - Failure!
Build: http://jenkins.thetaphi.de/job/Lucene-Solr-trunk-Linux/12488/
Java: 32bit/jdk1.8.0_60-ea-b12 -server -XX:+UseConcMarkSweepGC
1 tests failed.
FAILED: org.apache.solr.cloud.HttpPartitionTest.test
Error Message:
Didn't see all replicas for shard shard1 in c8n_1x2 come up within 30000 ms! ClusterState: { "collection1":{ "replicationFactor":"1", "shards":{ "shard1":{ "range":"80000000-ffffffff", "state":"active", "replicas":{"core_node2":{ "core":"collection1", "base_url":"http://127.0.0.1:36422/y_ow", "node_name":"127.0.0.1:36422_y_ow", "state":"active", "leader":"true"}}}, "shard2":{ "range":"0-7fffffff", "state":"active", "replicas":{ "core_node1":{ "core":"collection1", "base_url":"http://127.0.0.1:41627/y_ow", "node_name":"127.0.0.1:41627_y_ow", "state":"active", "leader":"true"}, "core_node3":{ "core":"collection1", "base_url":"http://127.0.0.1:35144/y_ow", "node_name":"127.0.0.1:35144_y_ow", "state":"active"}}}}, "router":{"name":"compositeId"}, "maxShardsPerNode":"1", "autoAddReplicas":"false", "autoCreated":"true"}, "control_collection":{ "replicationFactor":"1", "shards":{"shard1":{ "range":"80000000-7fffffff", "state":"active", "replicas":{"core_node1":{ "core":"collection1", "base_url":"http://127.0.0.1:50216/y_ow", "node_name":"127.0.0.1:50216_y_ow", "state":"active", "leader":"true"}}}}, "router":{"name":"compositeId"}, "maxShardsPerNode":"1", "autoAddReplicas":"false", "autoCreated":"true"}, "c8n_1x2":{ "replicationFactor":"2", "shards":{"shard1":{ "range":"80000000-7fffffff", "state":"active", "replicas":{ "core_node1":{ "core":"c8n_1x2_shard1_replica1", "base_url":"http://127.0.0.1:36422/y_ow", "node_name":"127.0.0.1:36422_y_ow", "state":"recovering"}, "core_node2":{ "core":"c8n_1x2_shard1_replica2", "base_url":"http://127.0.0.1:35144/y_ow", "node_name":"127.0.0.1:35144_y_ow", "state":"active", "leader":"true"}}}}, "router":{"name":"compositeId"}, "maxShardsPerNode":"1", "autoAddReplicas":"false"}}
Stack Trace:
java.lang.AssertionError: Didn't see all replicas for shard shard1 in c8n_1x2 come up within 30000 ms! ClusterState: {
"collection1":{
"replicationFactor":"1",
"shards":{
"shard1":{
"range":"80000000-ffffffff",
"state":"active",
"replicas":{"core_node2":{
"core":"collection1",
"base_url":"http://127.0.0.1:36422/y_ow",
"node_name":"127.0.0.1:36422_y_ow",
"state":"active",
"leader":"true"}}},
"shard2":{
"range":"0-7fffffff",
"state":"active",
"replicas":{
"core_node1":{
"core":"collection1",
"base_url":"http://127.0.0.1:41627/y_ow",
"node_name":"127.0.0.1:41627_y_ow",
"state":"active",
"leader":"true"},
"core_node3":{
"core":"collection1",
"base_url":"http://127.0.0.1:35144/y_ow",
"node_name":"127.0.0.1:35144_y_ow",
"state":"active"}}}},
"router":{"name":"compositeId"},
"maxShardsPerNode":"1",
"autoAddReplicas":"false",
"autoCreated":"true"},
"control_collection":{
"replicationFactor":"1",
"shards":{"shard1":{
"range":"80000000-7fffffff",
"state":"active",
"replicas":{"core_node1":{
"core":"collection1",
"base_url":"http://127.0.0.1:50216/y_ow",
"node_name":"127.0.0.1:50216_y_ow",
"state":"active",
"leader":"true"}}}},
"router":{"name":"compositeId"},
"maxShardsPerNode":"1",
"autoAddReplicas":"false",
"autoCreated":"true"},
"c8n_1x2":{
"replicationFactor":"2",
"shards":{"shard1":{
"range":"80000000-7fffffff",
"state":"active",
"replicas":{
"core_node1":{
"core":"c8n_1x2_shard1_replica1",
"base_url":"http://127.0.0.1:36422/y_ow",
"node_name":"127.0.0.1:36422_y_ow",
"state":"recovering"},
"core_node2":{
"core":"c8n_1x2_shard1_replica2",
"base_url":"http://127.0.0.1:35144/y_ow",
"node_name":"127.0.0.1:35144_y_ow",
"state":"active",
"leader":"true"}}}},
"router":{"name":"compositeId"},
"maxShardsPerNode":"1",
"autoAddReplicas":"false"}}
at __randomizedtesting.SeedInfo.seed([375FC4C9173DBE9E:BF0BFB13B9C1D366]:0)
at org.junit.Assert.fail(Assert.java:93)
at org.apache.solr.cloud.AbstractFullDistribZkTestBase.ensureAllReplicasAreActive(AbstractFullDistribZkTestBase.java:1920)
at org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:205)
at org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:105)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1627)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:836)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:872)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:886)
at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:960)
at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:935)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:50)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:49)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:798)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:458)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:845)
at com.carrotsearch.randomizedtesting.RandomizedRunner$3.evaluate(RandomizedRunner.java:747)
at com.carrotsearch.randomizedtesting.RandomizedRunner$4.evaluate(RandomizedRunner.java:781)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:792)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:46)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:42)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:39)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:54)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:48)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:65)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:55)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:365)
at java.lang.Thread.run(Thread.java:745)
Build Log:
[...truncated 1709 lines...]
[junit4] JVM J0: stdout was not empty, see: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/build/core/test/temp/junit4-J0-20150428_145157_996.sysout
[junit4] >>> JVM J0: stdout (verbatim) ----
[junit4] Default case invoked for:
[junit4] opcode = 0, "Node"
[junit4] Default case invoked for:
[junit4] opcode = 0, "Node"
[junit4] Default case invoked for:
[junit4] opcode = 199, "Phi"
[junit4] <<< JVM J0: EOF ----
[...truncated 8742 lines...]
[junit4] Suite: org.apache.solr.cloud.HttpPartitionTest
[junit4] 2> Creating dataDir: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/init-core-data-001
[junit4] 2> 960547 T6226 oas.BaseDistributedSearchTestCase.initHostContext Setting hostContext system property: /y_ow/
[junit4] 2> 960549 T6226 oasc.ZkTestServer.run STARTING ZK TEST SERVER
[junit4] 2> 960549 T6227 oasc.ZkTestServer$2$1.setClientPort client port:0.0.0.0/0.0.0.0:0
[junit4] 2> 960550 T6227 oasc.ZkTestServer$ZKServerMain.runFromConfig Starting server
[junit4] 2> 960649 T6226 oasc.ZkTestServer.run start zk server on port:43193
[junit4] 2> 960663 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig-tlog.xml to /configs/conf1/solrconfig.xml
[junit4] 2> 960664 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/schema.xml to /configs/conf1/schema.xml
[junit4] 2> 960665 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/solrconfig.snippet.randomindexconfig.xml to /configs/conf1/solrconfig.snippet.randomindexconfig.xml
[junit4] 2> 960666 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/stopwords.txt to /configs/conf1/stopwords.txt
[junit4] 2> 960667 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/protwords.txt to /configs/conf1/protwords.txt
[junit4] 2> 960668 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/currency.xml to /configs/conf1/currency.xml
[junit4] 2> 960669 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/enumsConfig.xml to /configs/conf1/enumsConfig.xml
[junit4] 2> 960669 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/open-exchange-rates.json to /configs/conf1/open-exchange-rates.json
[junit4] 2> 960670 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/mapping-ISOLatin1Accent.txt to /configs/conf1/mapping-ISOLatin1Accent.txt
[junit4] 2> 960671 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/old_synonyms.txt to /configs/conf1/old_synonyms.txt
[junit4] 2> 960672 T6226 oasc.AbstractZkTestCase.putConfig put /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/core/src/test-files/solr/collection1/conf/synonyms.txt to /configs/conf1/synonyms.txt
[junit4] 2> 960726 T6226 oas.SolrTestCaseJ4.writeCoreProperties Writing core.properties file to /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1
[junit4] 2> 960728 T6226 oejs.Server.doStart jetty-9.2.10.v20150310
[junit4] 2> 960729 T6226 oejsh.ContextHandler.doStart Started o.e.j.s.ServletContextHandler@1030194{/y_ow,null,AVAILABLE}
[junit4] 2> 960735 T6226 oejs.AbstractConnector.doStart Started ServerConnector@1aa997{HTTP/1.1}{127.0.0.1:40508}
[junit4] 2> 960735 T6226 oejs.Server.doStart Started @961672ms
[junit4] 2> 960735 T6226 oascse.JettySolrRunner$1.lifeCycleStarted Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/tempDir-001/control/data, hostContext=/y_ow, hostPort=50216, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores}
[junit4] 2> 960736 T6226 oass.SolrDispatchFilter.init SolrDispatchFilter.init()sun.misc.Launcher$AppClassLoader@e2f2a
[junit4] 2> 960736 T6226 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/'
[junit4] 2> 960747 T6226 oasc.SolrXmlConfig.fromFile Loading container configuration from /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/solr.xml
[junit4] 2> 960758 T6226 oasc.CorePropertiesLocator.<init> Config-defined core root directory: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores
[junit4] 2> 960759 T6226 oasc.CoreContainer.<init> New CoreContainer 2457809
[junit4] 2> 960759 T6226 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/]
[junit4] 2> 960760 T6226 oasc.CoreContainer.load loading shared library: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/lib
[junit4] 2> 960760 T6226 oasc.SolrResourceLoader.addToClassLoader WARN Can't find (or read) directory to add to classloader: lib (resolved as: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/lib).
[junit4] 2> 960764 T6226 oashc.HttpShardHandlerFactory.init created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
[junit4] 2> 960765 T6226 oasu.UpdateShardHandler.<init> Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
[junit4] 2> 960766 T6226 oasl.LogWatcher.createWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
[junit4] 2> 960766 T6226 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
[junit4] 2> 960766 T6226 oasc.CoreContainer.load Node Name: 127.0.0.1
[junit4] 2> 960766 T6226 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:43193/solr
[junit4] 2> 960767 T6226 oasc.ZkController.checkChrootPath zkHost includes chroot
[junit4] 2> 960774 T6226 n:127.0.0.1:50216_y_ow oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:50216_y_ow
[junit4] 2> 960776 T6226 n:127.0.0.1:50216_y_ow oasc.Overseer.close Overseer (id=null) closing
[junit4] 2> 960777 T6226 n:127.0.0.1:50216_y_ow oasc.OverseerElectionContext.runLeaderProcess I am going to be the leader 127.0.0.1:50216_y_ow
[junit4] 2> 960777 T6226 n:127.0.0.1:50216_y_ow oasc.Overseer.start Overseer (id=93731867006337027-127.0.0.1:50216_y_ow-n_0000000000) starting
[junit4] 2> 960780 T6226 n:127.0.0.1:50216_y_ow oasc.OverseerAutoReplicaFailoverThread.<init> Starting OverseerAutoReplicaFailoverThread autoReplicaFailoverWorkLoopDelay=10000 autoReplicaFailoverWaitAfterExpiration=30000 autoReplicaFailoverBadNodeExpiration=60000
[junit4] 2> 960781 T6256 n:127.0.0.1:50216_y_ow oasc.OverseerCollectionProcessor.run Process current queue of collection creations
[junit4] 2> 960781 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run Starting to work on the main queue
[junit4] 2> 960783 T6226 n:127.0.0.1:50216_y_ow oasc.CorePropertiesLocator.discover Looking for core definitions underneath /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores
[junit4] 2> 960784 T6226 n:127.0.0.1:50216_y_ow oasc.CoreDescriptor.<init> CORE DESCRIPTOR: {name=collection1, config=solrconfig.xml, transient=false, schema=schema.xml, loadOnStartup=true, instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1, collection=control_collection, absoluteInstDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/, coreNodeName=, dataDir=data/, shard=}
[junit4] 2> 960784 T6226 n:127.0.0.1:50216_y_ow oasc.CorePropertiesLocator.discoverUnder Found core collection1 in /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/
[junit4] 2> 960784 T6226 n:127.0.0.1:50216_y_ow oasc.CorePropertiesLocator.discover Found 1 core definitions
[junit4] 2> 960785 T6258 n:127.0.0.1:50216_y_ow c:control_collection x:collection1 oasc.ZkController.publish publishing core=collection1 state=down collection=control_collection
[junit4] 2> 960785 T6258 n:127.0.0.1:50216_y_ow c:control_collection x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
[junit4] 2> 960785 T6258 n:127.0.0.1:50216_y_ow oasc.ZkController.waitForCoreNodeName look for our core node name
[junit4] 2> 960785 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 960786 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "core":"collection1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:50216/y_ow",
[junit4] 2> "node_name":"127.0.0.1:50216_y_ow",
[junit4] 2> "numShards":"1",
[junit4] 2> "state":"down",
[junit4] 2> "shard":null,
[junit4] 2> "collection":"control_collection",
[junit4] 2> "operation":"state"} current state version: 0
[junit4] 2> 960786 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=1 message={
[junit4] 2> "core":"collection1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:50216/y_ow",
[junit4] 2> "node_name":"127.0.0.1:50216_y_ow",
[junit4] 2> "numShards":"1",
[junit4] 2> "state":"down",
[junit4] 2> "shard":null,
[junit4] 2> "collection":"control_collection",
[junit4] 2> "operation":"state"}
[junit4] 2> 960786 T6255 n:127.0.0.1:50216_y_ow oasco.ClusterStateMutator.createCollection building a new cName: control_collection
[junit4] 2> 960786 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Assigning new node to shard shard=shard1
[junit4] 2> 961785 T6258 n:127.0.0.1:50216_y_ow oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
[junit4] 2> 961786 T6258 n:127.0.0.1:50216_y_ow oasc.ZkController.createCollectionZkNode Check for collection zkNode:control_collection
[junit4] 2> 961786 T6258 n:127.0.0.1:50216_y_ow oasc.ZkController.createCollectionZkNode Collection zkNode exists
[junit4] 2> 961787 T6258 n:127.0.0.1:50216_y_ow oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/'
[junit4] 2> 961796 T6258 n:127.0.0.1:50216_y_ow oasc.Config.<init> loaded config solrconfig.xml with version 0
[junit4] 2> 961798 T6258 n:127.0.0.1:50216_y_ow oasc.SolrConfig.refreshRequestParams current version of requestparams : -1
[junit4] 2> 961801 T6258 n:127.0.0.1:50216_y_ow oasc.SolrConfig.<init> Using Lucene MatchVersion: 6.0.0
[junit4] 2> 961807 T6258 n:127.0.0.1:50216_y_ow oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
[junit4] 2> 961808 T6258 n:127.0.0.1:50216_y_ow oass.IndexSchema.readSchema Reading Solr Schema from /configs/conf1/schema.xml
[junit4] 2> 961812 T6258 n:127.0.0.1:50216_y_ow oass.IndexSchema.readSchema [collection1] Schema name=test
[junit4] 2> 961879 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
[junit4] 2> 961882 T6258 n:127.0.0.1:50216_y_ow oass.IndexSchema.readSchema default search field in schema is text
[junit4] 2> 961883 T6258 n:127.0.0.1:50216_y_ow oass.IndexSchema.readSchema unique key field: id
[junit4] 2> 961887 T6258 n:127.0.0.1:50216_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
[junit4] 2> 961889 T6258 n:127.0.0.1:50216_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
[junit4] 2> 961890 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
[junit4] 2> 961890 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
[junit4] 2> 961890 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
[junit4] 2> 961890 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
[junit4] 2> 961891 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
[junit4] 2> 961891 T6258 n:127.0.0.1:50216_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
[junit4] 2> 961891 T6258 n:127.0.0.1:50216_y_ow oasc.CoreContainer.create Creating SolrCore 'collection1' using configuration from collection control_collection
[junit4] 2> 961891 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
[junit4] 2> 961891 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.<init> [[collection1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/], dataDir=[null]
[junit4] 2> 961892 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a50e1b
[junit4] 2> 961892 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/data
[junit4] 2> 961892 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/data/index/
[junit4] 2> 961893 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.initIndex WARN [collection1] Solr index directory '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/data/index' doesn't exist. Creating new index...
[junit4] 2> 961893 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/control-001/cores/collection1/data/index
[junit4] 2> 961893 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=29, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.18017383033454096]
[junit4] 2> 961894 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
[junit4] 2> commit{dir=MockDirectoryWrapper(RAMDirectory@880570 lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@1f9bd62),segFN=segments_1,generation=1}
[junit4] 2> 961894 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
[junit4] 2> 961899 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "nodistrib"
[junit4] 2> 961899 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "dedupe"
[junit4] 2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
[junit4] 2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "stored_sig"
[junit4] 2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
[junit4] 2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-explicit"
[junit4] 2> 961900 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-implicit"
[junit4] 2> 961901 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
[junit4] 2> 961901 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.loadUpdateProcessorChains no updateRequestProcessorChain defined as default, creating implicit default
[junit4] 2> 961903 T6258 n:127.0.0.1:50216_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 961904 T6258 n:127.0.0.1:50216_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 961904 T6258 n:127.0.0.1:50216_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 961905 T6258 n:127.0.0.1:50216_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 961908 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.RequestHandlers.initHandlersFromConfig Registered paths: /admin/mbeans,standard,/update/csv,/update/json/docs,/admin/luke,/admin/segments,/get,/admin/system,/replication,/admin/properties,/config,/schema,/admin/plugins,/admin/logging,/update/json,/admin/threads,/admin/ping,/update,/admin/file
[junit4] 2> 961908 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.initStatsCache Using default statsCache cache: org.apache.solr.search.stats.LocalStatsCache
[junit4] 2> 961909 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.UpdateHandler.<init> Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 961909 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.UpdateLog.init Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10
[junit4] 2> 961909 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.CommitTracker.<init> Hard AutoCommit: disabled
[junit4] 2> 961910 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.CommitTracker.<init> Soft AutoCommit: disabled
[junit4] 2> 961910 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=46, maxMergeAtOnceExplicit=46, maxMergedSegmentMB=82.8984375, floorSegmentMB=1.099609375, forceMergeDeletesPctAllowed=15.853958222535056, segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
[junit4] 2> 961910 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
[junit4] 2> commit{dir=MockDirectoryWrapper(RAMDirectory@880570 lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@1f9bd62),segFN=segments_1,generation=1}
[junit4] 2> 961911 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
[junit4] 2> 961911 T6258 n:127.0.0.1:50216_y_ow x:collection1 oass.SolrIndexSearcher.<init> Opening Searcher@1f42150[collection1] main
[junit4] 2> 961911 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage.newStorageIO Setting up ZooKeeper-based storage for the RestManager with znodeBase: /configs/conf1
[junit4] 2> 961912 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.configure Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 961912 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.RestManager.init Initializing RestManager with initArgs: {}
[junit4] 2> 961912 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage.load Reading _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 961912 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.openInputStream No data found for znode /configs/conf1/_rest_managed.json
[junit4] 2> 961913 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.ManagedResourceStorage.load Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 961913 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasr.RestManager.init Initializing 0 registered ManagedResources
[junit4] 2> 961913 T6258 n:127.0.0.1:50216_y_ow x:collection1 oash.ReplicationHandler.inform Commits will be reserved for 10000
[junit4] 2> 961913 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.ZkController.getConfDirListeners watch zkdir /configs/conf1
[junit4] 2> 961913 T6259 n:127.0.0.1:50216_y_ow x:collection1 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@1f42150[collection1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 961914 T6258 n:127.0.0.1:50216_y_ow x:collection1 oasc.CoreContainer.registerCore registering core: collection1
[junit4] 2> 961914 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:50216/y_ow collection:control_collection shard:shard1
[junit4] 2> 961914 T6226 n:127.0.0.1:50216_y_ow oass.SolrDispatchFilter.init user.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0
[junit4] 2> 961915 T6226 n:127.0.0.1:50216_y_ow oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
[junit4] 2> 961918 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
[junit4] 2> 961923 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 961923 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
[junit4] 2> 961923 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
[junit4] 2> 961923 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "operation":"leader",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection"} current state version: 1
[junit4] 2> ASYNC NEW_CORE C5782 name=collection1 org.apache.solr.core.SolrCore@5387a6 url=http://127.0.0.1:50216/y_ow/collection1 node=127.0.0.1:50216_y_ow C5782_STATE=coll:control_collection core:collection1 props:{core=collection1, base_url=http://127.0.0.1:50216/y_ow, node_name=127.0.0.1:50216_y_ow, state=down}
[junit4] 2> 961923 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 C5782 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:50216/y_ow/collection1/
[junit4] 2> 961924 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 C5782 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
[junit4] 2> 961924 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 C5782 oasc.SyncStrategy.syncToMe http://127.0.0.1:50216/y_ow/collection1/ has no replicas
[junit4] 2> 961923 T6226 oasc.ChaosMonkey.monkeyLog monkey: init - expire sessions:false cause connection loss:false
[junit4] 2> 961924 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:50216/y_ow/collection1/ shard1
[junit4] 2> 961928 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 961928 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "operation":"leader",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection",
[junit4] 2> "base_url":"http://127.0.0.1:50216/y_ow",
[junit4] 2> "core":"collection1",
[junit4] 2> "state":"active"} current state version: 1
[junit4] 2> 961982 T6226 oas.SolrTestCaseJ4.writeCoreProperties Writing core.properties file to /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1
[junit4] 2> 961983 T6226 oasc.AbstractFullDistribZkTestBase.createJettys create jetty 1 in directory /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001
[junit4] 2> 961984 T6226 oejs.Server.doStart jetty-9.2.10.v20150310
[junit4] 2> 961985 T6226 oejsh.ContextHandler.doStart Started o.e.j.s.ServletContextHandler@df06d0{/y_ow,null,AVAILABLE}
[junit4] 2> 961985 T6226 oejs.AbstractConnector.doStart Started ServerConnector@1cebdab{HTTP/1.1}{127.0.0.1:35724}
[junit4] 2> 961985 T6226 oejs.Server.doStart Started @962922ms
[junit4] 2> 961986 T6226 oascse.JettySolrRunner$1.lifeCycleStarted Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/tempDir-001/jetty1, solrconfig=solrconfig.xml, hostContext=/y_ow, hostPort=41627, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores}
[junit4] 2> 961986 T6226 oass.SolrDispatchFilter.init SolrDispatchFilter.init()sun.misc.Launcher$AppClassLoader@e2f2a
[junit4] 2> 961986 T6226 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/'
[junit4] 2> 961995 T6226 oasc.SolrXmlConfig.fromFile Loading container configuration from /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/solr.xml
[junit4] 2> 961999 T6226 oasc.CorePropertiesLocator.<init> Config-defined core root directory: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores
[junit4] 2> 962000 T6226 oasc.CoreContainer.<init> New CoreContainer 5753138
[junit4] 2> 962000 T6226 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/]
[junit4] 2> 962000 T6226 oasc.CoreContainer.load loading shared library: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/lib
[junit4] 2> 962000 T6226 oasc.SolrResourceLoader.addToClassLoader WARN Can't find (or read) directory to add to classloader: lib (resolved as: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/lib).
[junit4] 2> 962004 T6226 oashc.HttpShardHandlerFactory.init created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
[junit4] 2> 962005 T6226 oasu.UpdateShardHandler.<init> Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
[junit4] 2> 962006 T6226 oasl.LogWatcher.createWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
[junit4] 2> 962006 T6226 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
[junit4] 2> 962006 T6226 oasc.CoreContainer.load Node Name: 127.0.0.1
[junit4] 2> 962006 T6226 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:43193/solr
[junit4] 2> 962007 T6226 oasc.ZkController.checkChrootPath zkHost includes chroot
[junit4] 2> 962078 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.register We are http://127.0.0.1:50216/y_ow/collection1/ and leader is http://127.0.0.1:50216/y_ow/collection1/
[junit4] 2> 962079 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:50216/y_ow
[junit4] 2> 962079 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
[junit4] 2> 962079 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.publish publishing core=collection1 state=active collection=control_collection
[junit4] 2> 962079 T6262 n:127.0.0.1:50216_y_ow c:control_collection s:shard1 x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
[junit4] 2> 962080 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 962080 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "core":"collection1",
[junit4] 2> "core_node_name":"core_node1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:50216/y_ow",
[junit4] 2> "node_name":"127.0.0.1:50216_y_ow",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"active",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection",
[junit4] 2> "operation":"state"} current state version: 2
[junit4] 2> 962081 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=2 message={
[junit4] 2> "core":"collection1",
[junit4] 2> "core_node_name":"core_node1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:50216/y_ow",
[junit4] 2> "node_name":"127.0.0.1:50216_y_ow",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"active",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"control_collection",
[junit4] 2> "operation":"state"}
[junit4] 2> 963013 T6226 n:127.0.0.1:41627_y_ow oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:41627_y_ow
[junit4] 2> 963015 T6226 n:127.0.0.1:41627_y_ow oasc.Overseer.close Overseer (id=null) closing
[junit4] 2> 963016 T6226 n:127.0.0.1:41627_y_ow oasc.CorePropertiesLocator.discover Looking for core definitions underneath /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores
[junit4] 2> 963017 T6226 n:127.0.0.1:41627_y_ow oasc.CoreDescriptor.<init> CORE DESCRIPTOR: {name=collection1, config=solrconfig.xml, transient=false, schema=schema.xml, loadOnStartup=true, instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1, collection=collection1, absoluteInstDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/, coreNodeName=, dataDir=data/, shard=}
[junit4] 2> 963017 T6226 n:127.0.0.1:41627_y_ow oasc.CorePropertiesLocator.discoverUnder Found core collection1 in /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/
[junit4] 2> 963018 T6226 n:127.0.0.1:41627_y_ow oasc.CorePropertiesLocator.discover Found 1 core definitions
[junit4] 2> 963019 T6284 n:127.0.0.1:41627_y_ow c:collection1 x:collection1 oasc.ZkController.publish publishing core=collection1 state=down collection=collection1
[junit4] 2> 963019 T6284 n:127.0.0.1:41627_y_ow c:collection1 x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
[junit4] 2> 963019 T6284 n:127.0.0.1:41627_y_ow oasc.ZkController.waitForCoreNodeName look for our core node name
[junit4] 2> 963019 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 963020 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "core":"collection1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:41627/y_ow",
[junit4] 2> "node_name":"127.0.0.1:41627_y_ow",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"down",
[junit4] 2> "shard":null,
[junit4] 2> "collection":"collection1",
[junit4] 2> "operation":"state"} current state version: 3
[junit4] 2> 963020 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=2 message={
[junit4] 2> "core":"collection1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:41627/y_ow",
[junit4] 2> "node_name":"127.0.0.1:41627_y_ow",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"down",
[junit4] 2> "shard":null,
[junit4] 2> "collection":"collection1",
[junit4] 2> "operation":"state"}
[junit4] 2> 963021 T6255 n:127.0.0.1:50216_y_ow oasco.ClusterStateMutator.createCollection building a new cName: collection1
[junit4] 2> 963021 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Assigning new node to shard shard=shard2
[junit4] 2> 964020 T6284 n:127.0.0.1:41627_y_ow oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
[junit4] 2> 964020 T6284 n:127.0.0.1:41627_y_ow oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
[junit4] 2> 964020 T6284 n:127.0.0.1:41627_y_ow oasc.ZkController.createCollectionZkNode Collection zkNode exists
[junit4] 2> 964021 T6284 n:127.0.0.1:41627_y_ow oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/'
[junit4] 2> 964033 T6284 n:127.0.0.1:41627_y_ow oasc.Config.<init> loaded config solrconfig.xml with version 0
[junit4] 2> 964038 T6284 n:127.0.0.1:41627_y_ow oasc.SolrConfig.refreshRequestParams current version of requestparams : -1
[junit4] 2> 964041 T6284 n:127.0.0.1:41627_y_ow oasc.SolrConfig.<init> Using Lucene MatchVersion: 6.0.0
[junit4] 2> 964068 T6284 n:127.0.0.1:41627_y_ow oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
[junit4] 2> 964072 T6284 n:127.0.0.1:41627_y_ow oass.IndexSchema.readSchema Reading Solr Schema from /configs/conf1/schema.xml
[junit4] 2> 964076 T6284 n:127.0.0.1:41627_y_ow oass.IndexSchema.readSchema [collection1] Schema name=test
[junit4] 2> 964149 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
[junit4] 2> 964154 T6284 n:127.0.0.1:41627_y_ow oass.IndexSchema.readSchema default search field in schema is text
[junit4] 2> 964155 T6284 n:127.0.0.1:41627_y_ow oass.IndexSchema.readSchema unique key field: id
[junit4] 2> 964159 T6284 n:127.0.0.1:41627_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
[junit4] 2> 964160 T6284 n:127.0.0.1:41627_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
[junit4] 2> 964161 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
[junit4] 2> 964162 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
[junit4] 2> 964162 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
[junit4] 2> 964162 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
[junit4] 2> 964164 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
[junit4] 2> 964165 T6284 n:127.0.0.1:41627_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
[junit4] 2> 964165 T6284 n:127.0.0.1:41627_y_ow oasc.CoreContainer.create Creating SolrCore 'collection1' using configuration from collection collection1
[junit4] 2> 964165 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
[junit4] 2> 964165 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.<init> [[collection1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/], dataDir=[null]
[junit4] 2> 964165 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a50e1b
[junit4] 2> 964166 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/data
[junit4] 2> 964166 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/data/index/
[junit4] 2> 964166 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.initIndex WARN [collection1] Solr index directory '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/data/index' doesn't exist. Creating new index...
[junit4] 2> 964166 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-1-001/cores/collection1/data/index
[junit4] 2> 964167 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=29, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.18017383033454096]
[junit4] 2> 964167 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
[junit4] 2> commit{dir=MockDirectoryWrapper(RAMDirectory@b4ca8f lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@4bb787),segFN=segments_1,generation=1}
[junit4] 2> 964167 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
[junit4] 2> 964169 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "nodistrib"
[junit4] 2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "dedupe"
[junit4] 2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
[junit4] 2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "stored_sig"
[junit4] 2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
[junit4] 2> 964170 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-explicit"
[junit4] 2> 964171 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-implicit"
[junit4] 2> 964171 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
[junit4] 2> 964171 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.loadUpdateProcessorChains no updateRequestProcessorChain defined as default, creating implicit default
[junit4] 2> 964172 T6284 n:127.0.0.1:41627_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 964173 T6284 n:127.0.0.1:41627_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 964174 T6284 n:127.0.0.1:41627_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 964174 T6284 n:127.0.0.1:41627_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 964178 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.RequestHandlers.initHandlersFromConfig Registered paths: /admin/mbeans,standard,/update/csv,/update/json/docs,/admin/luke,/admin/segments,/get,/admin/system,/replication,/admin/properties,/config,/schema,/admin/plugins,/admin/logging,/update/json,/admin/threads,/admin/ping,/update,/admin/file
[junit4] 2> 964179 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.initStatsCache Using default statsCache cache: org.apache.solr.search.stats.LocalStatsCache
[junit4] 2> 964179 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.UpdateHandler.<init> Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 964179 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.UpdateLog.init Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10
[junit4] 2> 964180 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.CommitTracker.<init> Hard AutoCommit: disabled
[junit4] 2> 964180 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.CommitTracker.<init> Soft AutoCommit: disabled
[junit4] 2> 964181 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=46, maxMergeAtOnceExplicit=46, maxMergedSegmentMB=82.8984375, floorSegmentMB=1.099609375, forceMergeDeletesPctAllowed=15.853958222535056, segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
[junit4] 2> 964181 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
[junit4] 2> commit{dir=MockDirectoryWrapper(RAMDirectory@b4ca8f lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@4bb787),segFN=segments_1,generation=1}
[junit4] 2> 964181 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
[junit4] 2> 964181 T6284 n:127.0.0.1:41627_y_ow x:collection1 oass.SolrIndexSearcher.<init> Opening Searcher@1f712d6[collection1] main
[junit4] 2> 964182 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage.newStorageIO Setting up ZooKeeper-based storage for the RestManager with znodeBase: /configs/conf1
[junit4] 2> 964182 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.configure Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 964182 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.RestManager.init Initializing RestManager with initArgs: {}
[junit4] 2> 964182 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage.load Reading _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 964183 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.openInputStream No data found for znode /configs/conf1/_rest_managed.json
[junit4] 2> 964183 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.ManagedResourceStorage.load Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 964183 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasr.RestManager.init Initializing 0 registered ManagedResources
[junit4] 2> 964183 T6284 n:127.0.0.1:41627_y_ow x:collection1 oash.ReplicationHandler.inform Commits will be reserved for 10000
[junit4] 2> 964184 T6285 n:127.0.0.1:41627_y_ow x:collection1 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@1f712d6[collection1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 964184 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.ZkController.getConfDirListeners watch zkdir /configs/conf1
[junit4] 2> 964184 T6284 n:127.0.0.1:41627_y_ow x:collection1 oasc.CoreContainer.registerCore registering core: collection1
[junit4] 2> 964200 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:41627/y_ow collection:collection1 shard:shard2
[junit4] 2> 964200 T6226 n:127.0.0.1:41627_y_ow oass.SolrDispatchFilter.init user.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0
[junit4] 2> 964200 T6226 n:127.0.0.1:41627_y_ow oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
[junit4] 2> 964203 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard2
[junit4] 2> 964204 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
[junit4] 2> 964204 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
[junit4] 2> ASYNC NEW_CORE C5783 name=collection1 org.apache.solr.core.SolrCore@6da5c2 url=http://127.0.0.1:41627/y_ow/collection1 node=127.0.0.1:41627_y_ow C5783_STATE=coll:collection1 core:collection1 props:{core=collection1, base_url=http://127.0.0.1:41627/y_ow, node_name=127.0.0.1:41627_y_ow, state=down}
[junit4] 2> 964204 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 C5783 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:41627/y_ow/collection1/
[junit4] 2> 964204 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 C5783 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
[junit4] 2> 964205 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 C5783 oasc.SyncStrategy.syncToMe http://127.0.0.1:41627/y_ow/collection1/ has no replicas
[junit4] 2> 964205 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:41627/y_ow/collection1/ shard2
[junit4] 2> 964206 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 964207 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 2, message = {
[junit4] 2> "operation":"leader",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1"} current state version: 4
[junit4] 2> 964211 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "operation":"leader",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1",
[junit4] 2> "base_url":"http://127.0.0.1:41627/y_ow",
[junit4] 2> "core":"collection1",
[junit4] 2> "state":"active"} current state version: 4
[junit4] 2> 964212 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 964256 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.register We are http://127.0.0.1:41627/y_ow/collection1/ and leader is http://127.0.0.1:41627/y_ow/collection1/
[junit4] 2> 964257 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.register No LogReplay needed for core=collection1 baseURL=http://127.0.0.1:41627/y_ow
[junit4] 2> 964257 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.checkRecovery I am the leader, no recovery necessary
[junit4] 2> 964257 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.publish publishing core=collection1 state=active collection=collection1
[junit4] 2> 964257 T6288 n:127.0.0.1:41627_y_ow c:collection1 s:shard2 x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
[junit4] 2> 964258 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 964263 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "core":"collection1",
[junit4] 2> "core_node_name":"core_node1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:41627/y_ow",
[junit4] 2> "node_name":"127.0.0.1:41627_y_ow",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"active",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1",
[junit4] 2> "operation":"state"} current state version: 5
[junit4] 2> 964265 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=2 message={
[junit4] 2> "core":"collection1",
[junit4] 2> "core_node_name":"core_node1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:41627/y_ow",
[junit4] 2> "node_name":"127.0.0.1:41627_y_ow",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"active",
[junit4] 2> "shard":"shard2",
[junit4] 2> "collection":"collection1",
[junit4] 2> "operation":"state"}
[junit4] 2> 964276 T6226 oas.SolrTestCaseJ4.writeCoreProperties Writing core.properties file to /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1
[junit4] 2> 964277 T6226 oasc.AbstractFullDistribZkTestBase.createJettys create jetty 2 in directory /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001
[junit4] 2> 964278 T6226 oejs.Server.doStart jetty-9.2.10.v20150310
[junit4] 2> 964279 T6226 oejsh.ContextHandler.doStart Started o.e.j.s.ServletContextHandler@bb9abb{/y_ow,null,AVAILABLE}
[junit4] 2> 964279 T6226 oejs.AbstractConnector.doStart Started ServerConnector@16435ad{HTTP/1.1}{127.0.0.1:40120}
[junit4] 2> 964280 T6226 oejs.Server.doStart Started @965216ms
[junit4] 2> 964280 T6226 oascse.JettySolrRunner$1.lifeCycleStarted Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/tempDir-001/jetty2, solrconfig=solrconfig.xml, hostContext=/y_ow, hostPort=36422, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores}
[junit4] 2> 964280 T6226 oass.SolrDispatchFilter.init SolrDispatchFilter.init()sun.misc.Launcher$AppClassLoader@e2f2a
[junit4] 2> 964281 T6226 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/'
[junit4] 2> 964291 T6226 oasc.SolrXmlConfig.fromFile Loading container configuration from /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/solr.xml
[junit4] 2> 964300 T6226 oasc.CorePropertiesLocator.<init> Config-defined core root directory: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores
[junit4] 2> 964300 T6226 oasc.CoreContainer.<init> New CoreContainer 12242390
[junit4] 2> 964301 T6226 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/]
[junit4] 2> 964301 T6226 oasc.CoreContainer.load loading shared library: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/lib
[junit4] 2> 964301 T6226 oasc.SolrResourceLoader.addToClassLoader WARN Can't find (or read) directory to add to classloader: lib (resolved as: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/lib).
[junit4] 2> 964306 T6226 oashc.HttpShardHandlerFactory.init created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
[junit4] 2> 964308 T6226 oasu.UpdateShardHandler.<init> Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
[junit4] 2> 964309 T6226 oasl.LogWatcher.createWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
[junit4] 2> 964309 T6226 oasl.LogWatcher.newRegisteredLogWatcher Registering Log Listener [Log4j (org.slf4j.impl.Log4jLoggerFactory)]
[junit4] 2> 964310 T6226 oasc.CoreContainer.load Node Name: 127.0.0.1
[junit4] 2> 964310 T6226 oasc.ZkContainer.initZooKeeper Zookeeper client=127.0.0.1:43193/solr
[junit4] 2> 964310 T6226 oasc.ZkController.checkChrootPath zkHost includes chroot
[junit4] 2> 965325 T6226 n:127.0.0.1:36422_y_ow oasc.ZkController.createEphemeralLiveNode Register node as live in ZooKeeper:/live_nodes/127.0.0.1:36422_y_ow
[junit4] 2> 965327 T6226 n:127.0.0.1:36422_y_ow oasc.Overseer.close Overseer (id=null) closing
[junit4] 2> 965328 T6226 n:127.0.0.1:36422_y_ow oasc.CorePropertiesLocator.discover Looking for core definitions underneath /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores
[junit4] 2> 965329 T6226 n:127.0.0.1:36422_y_ow oasc.CoreDescriptor.<init> CORE DESCRIPTOR: {name=collection1, config=solrconfig.xml, transient=false, schema=schema.xml, loadOnStartup=true, instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1, collection=collection1, absoluteInstDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/, coreNodeName=, dataDir=data/, shard=}
[junit4] 2> 965329 T6226 n:127.0.0.1:36422_y_ow oasc.CorePropertiesLocator.discoverUnder Found core collection1 in /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/
[junit4] 2> 965330 T6226 n:127.0.0.1:36422_y_ow oasc.CorePropertiesLocator.discover Found 1 core definitions
[junit4] 2> 965331 T6307 n:127.0.0.1:36422_y_ow c:collection1 x:collection1 oasc.ZkController.publish publishing core=collection1 state=down collection=collection1
[junit4] 2> 965332 T6307 n:127.0.0.1:36422_y_ow c:collection1 x:collection1 oasc.ZkController.publish numShards not found on descriptor - reading it from system property
[junit4] 2> 965333 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 965333 T6307 n:127.0.0.1:36422_y_ow oasc.ZkController.waitForCoreNodeName look for our core node name
[junit4] 2> 965333 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "core":"collection1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:36422/y_ow",
[junit4] 2> "node_name":"127.0.0.1:36422_y_ow",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"down",
[junit4] 2> "shard":null,
[junit4] 2> "collection":"collection1",
[junit4] 2> "operation":"state"} current state version: 6
[junit4] 2> 965334 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Update state numShards=2 message={
[junit4] 2> "core":"collection1",
[junit4] 2> "roles":null,
[junit4] 2> "base_url":"http://127.0.0.1:36422/y_ow",
[junit4] 2> "node_name":"127.0.0.1:36422_y_ow",
[junit4] 2> "numShards":"2",
[junit4] 2> "state":"down",
[junit4] 2> "shard":null,
[junit4] 2> "collection":"collection1",
[junit4] 2> "operation":"state"}
[junit4] 2> 965334 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Collection already exists with numShards=2
[junit4] 2> 965334 T6255 n:127.0.0.1:50216_y_ow oasco.ReplicaMutator.updateState Assigning new node to shard shard=shard1
[junit4] 2> 966333 T6307 n:127.0.0.1:36422_y_ow oasc.ZkController.waitForShardId waiting to find shard id in clusterstate for collection1
[junit4] 2> 966333 T6307 n:127.0.0.1:36422_y_ow oasc.ZkController.createCollectionZkNode Check for collection zkNode:collection1
[junit4] 2> 966334 T6307 n:127.0.0.1:36422_y_ow oasc.ZkController.createCollectionZkNode Collection zkNode exists
[junit4] 2> 966334 T6307 n:127.0.0.1:36422_y_ow oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/'
[junit4] 2> 966342 T6307 n:127.0.0.1:36422_y_ow oasc.Config.<init> loaded config solrconfig.xml with version 0
[junit4] 2> 966345 T6307 n:127.0.0.1:36422_y_ow oasc.SolrConfig.refreshRequestParams current version of requestparams : -1
[junit4] 2> 966348 T6307 n:127.0.0.1:36422_y_ow oasc.SolrConfig.<init> Using Lucene MatchVersion: 6.0.0
[junit4] 2> 966359 T6307 n:127.0.0.1:36422_y_ow oasc.SolrConfig.<init> Loaded SolrConfig: solrconfig.xml
[junit4] 2> 966360 T6307 n:127.0.0.1:36422_y_ow oass.IndexSchema.readSchema Reading Solr Schema from /configs/conf1/schema.xml
[junit4] 2> 966363 T6307 n:127.0.0.1:36422_y_ow oass.IndexSchema.readSchema [collection1] Schema name=test
[junit4] 2> 966462 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider.init Initialized with rates=open-exchange-rates.json, refreshInterval=1440.
[junit4] 2> 966467 T6307 n:127.0.0.1:36422_y_ow oass.IndexSchema.readSchema default search field in schema is text
[junit4] 2> 966468 T6307 n:127.0.0.1:36422_y_ow oass.IndexSchema.readSchema unique key field: id
[junit4] 2> 966472 T6307 n:127.0.0.1:36422_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
[junit4] 2> 966474 T6307 n:127.0.0.1:36422_y_ow oass.FileExchangeRateProvider.reload Reloading exchange rates from file currency.xml
[junit4] 2> 966475 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
[junit4] 2> 966475 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
[junit4] 2> 966476 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
[junit4] 2> 966476 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider.reload Reloading exchange rates from open-exchange-rates.json
[junit4] 2> 966476 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Unknown key IMPORTANT NOTE
[junit4] 2> 966476 T6307 n:127.0.0.1:36422_y_ow oass.OpenExchangeRatesOrgProvider$OpenExchangeRates.<init> WARN Expected key, got STRING
[junit4] 2> 966476 T6307 n:127.0.0.1:36422_y_ow oasc.CoreContainer.create Creating SolrCore 'collection1' using configuration from collection collection1
[junit4] 2> 966477 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.initDirectoryFactory org.apache.solr.core.MockDirectoryFactory
[junit4] 2> 966477 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.<init> [[collection1] ] Opening new SolrCore at [/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/], dataDir=[null]
[junit4] 2> 966477 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.JmxMonitoredMap.<init> JMX monitoring is enabled. Adding Solr mbeans to JMX Server: com.sun.jmx.mbeanserver.JmxMBeanServer@1a50e1b
[junit4] 2> 966477 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/data
[junit4] 2> 966478 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.getNewIndexDir New index directory detected: old=null new=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/data/index/
[junit4] 2> 966478 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.initIndex WARN [collection1] Solr index directory '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/data/index' doesn't exist. Creating new index...
[junit4] 2> 966478 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.CachingDirectoryFactory.get return new directory for /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-2-001/cores/collection1/data/index
[junit4] 2> 966478 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.LogByteSizeMergePolicy: [LogByteSizeMergePolicy: minMergeSize=1677721, mergeFactor=29, maxMergeSize=2147483648, maxMergeSizeForForcedMerge=9223372036854775807, calibrateSizeByDeletes=true, maxMergeDocs=2147483647, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=0.18017383033454096]
[junit4] 2> 966479 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrDeletionPolicy.onCommit SolrDeletionPolicy.onCommit: commits: num=1
[junit4] 2> commit{dir=MockDirectoryWrapper(RAMDirectory@c5953b lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@e6eed5),segFN=segments_1,generation=1}
[junit4] 2> 966479 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
[junit4] 2> 966481 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "nodistrib"
[junit4] 2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "dedupe"
[junit4] 2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "dedupe"
[junit4] 2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "stored_sig"
[junit4] 2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "stored_sig"
[junit4] 2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-explicit"
[junit4] 2> 966482 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init creating updateRequestProcessorChain "distrib-dup-test-chain-implicit"
[junit4] 2> 966483 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasup.UpdateRequestProcessorChain.init inserting DistributedUpdateProcessorFactory into updateRequestProcessorChain "distrib-dup-test-chain-implicit"
[junit4] 2> 966483 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.loadUpdateProcessorChains no updateRequestProcessorChain defined as default, creating implicit default
[junit4] 2> 966484 T6307 n:127.0.0.1:36422_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 966485 T6307 n:127.0.0.1:36422_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 966485 T6307 n:127.0.0.1:36422_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 966486 T6307 n:127.0.0.1:36422_y_ow x:collection1 oashl.XMLLoader.init xsltCacheLifetimeSeconds=60
[junit4] 2> 966489 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.RequestHandlers.initHandlersFromConfig Registered paths: /admin/mbeans,standard,/update/csv,/update/json/docs,/admin/luke,/admin/segments,/get,/admin/system,/replication,/admin/properties,/config,/schema,/admin/plugins,/admin/logging,/update/json,/admin/threads,/admin/ping,/update,/admin/file
[junit4] 2> 966489 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.initStatsCache Using default statsCache cache: org.apache.solr.search.stats.LocalStatsCache
[junit4] 2> 966490 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.UpdateHandler.<init> Using UpdateLog implementation: org.apache.solr.update.UpdateLog
[junit4] 2> 966490 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.UpdateLog.init Initializing UpdateLog: dataDir= defaultSyncLevel=FLUSH numRecordsToKeep=100 maxNumLogsToKeep=10
[junit4] 2> 966490 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.CommitTracker.<init> Hard AutoCommit: disabled
[junit4] 2> 966491 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.CommitTracker.<init> Soft AutoCommit: disabled
[junit4] 2> 966492 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasu.RandomMergePolicy.<init> RandomMergePolicy wrapping class org.apache.lucene.index.TieredMergePolicy: [TieredMergePolicy: maxMergeAtOnce=46, maxMergeAtOnceExplicit=46, maxMergedSegmentMB=82.8984375, floorSegmentMB=1.099609375, forceMergeDeletesPctAllowed=15.853958222535056, segmentsPerTier=20.0, maxCFSSegmentSizeMB=8.796093022207999E12, noCFSRatio=1.0
[junit4] 2> 966492 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrDeletionPolicy.onInit SolrDeletionPolicy.onInit: commits: num=1
[junit4] 2> commit{dir=MockDirectoryWrapper(RAMDirectory@c5953b lockFactory=org.apache.lucene.store.SingleInstanceLockFactory@e6eed5),segFN=segments_1,generation=1}
[junit4] 2> 966492 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrDeletionPolicy.updateCommits newest commit generation = 1
[junit4] 2> 966492 T6307 n:127.0.0.1:36422_y_ow x:collection1 oass.SolrIndexSearcher.<init> Opening Searcher@1feb4da[collection1] main
[junit4] 2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage.newStorageIO Setting up ZooKeeper-based storage for the RestManager with znodeBase: /configs/conf1
[junit4] 2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.configure Configured ZooKeeperStorageIO with znodeBase: /configs/conf1
[junit4] 2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.RestManager.init Initializing RestManager with initArgs: {}
[junit4] 2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage.load Reading _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 966493 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage$ZooKeeperStorageIO.openInputStream No data found for znode /configs/conf1/_rest_managed.json
[junit4] 2> 966494 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.ManagedResourceStorage.load Loaded null at path _rest_managed.json using ZooKeeperStorageIO:path=/configs/conf1
[junit4] 2> 966494 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasr.RestManager.init Initializing 0 registered ManagedResources
[junit4] 2> 966494 T6307 n:127.0.0.1:36422_y_ow x:collection1 oash.ReplicationHandler.inform Commits will be reserved for 10000
[junit4] 2> 966494 T6308 n:127.0.0.1:36422_y_ow x:collection1 oasc.SolrCore.registerSearcher [collection1] Registered new searcher Searcher@1feb4da[collection1] main{ExitableDirectoryReader(UninvertingDirectoryReader())}
[junit4] 2> 966495 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.ZkController.getConfDirListeners watch zkdir /configs/conf1
[junit4] 2> 966495 T6307 n:127.0.0.1:36422_y_ow x:collection1 oasc.CoreContainer.registerCore registering core: collection1
[junit4] 2> 966495 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ZkController.register Register replica - core:collection1 address:http://127.0.0.1:36422/y_ow collection:collection1 shard:shard1
[junit4] 2> 966495 T6226 n:127.0.0.1:36422_y_ow oass.SolrDispatchFilter.init user.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0
[junit4] 2> 966496 T6226 n:127.0.0.1:36422_y_ow oass.SolrDispatchFilter.init SolrDispatchFilter.init() done
[junit4] 2> 966497 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess Running the leader process for shard shard1
[junit4] 2> 966498 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 966498 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ShardLeaderElectionContext.waitForReplicasToComeUp Enough replicas found to continue.
[junit4] 2> 966498 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I may be the new leader - try and sync
[junit4] 2> ASYNC NEW_CORE C5784 name=collection1 org.apache.solr.core.SolrCore@d15410 url=http://127.0.0.1:36422/y_ow/collection1 node=127.0.0.1:36422_y_ow C5784_STATE=coll:collection1 core:collection1 props:{core=collection1, base_url=http://127.0.0.1:36422/y_ow, node_name=127.0.0.1:36422_y_ow, state=down}
[junit4] 2> 966498 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 C5784 oasc.SyncStrategy.sync Sync replicas to http://127.0.0.1:36422/y_ow/collection1/
[junit4] 2> 966498 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "operation":"leader",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"collection1"} current state version: 7
[junit4] 2> 966498 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 C5784 oasc.SyncStrategy.syncReplicas Sync Success - now sync replicas to me
[junit4] 2> 966499 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 C5784 oasc.SyncStrategy.syncToMe http://127.0.0.1:36422/y_ow/collection1/ has no replicas
[junit4] 2> 966499 T6311 n:127.0.0.1:36422_y_ow c:collection1 s:shard1 x:collection1 oasc.ShardLeaderElectionContext.runLeaderProcess I am the new leader: http://127.0.0.1:36422/y_ow/collection1/ shard1
[junit4] 2> 966500 T6254 n:127.0.0.1:50216_y_ow oasc.DistributedQueue$LatchWatcher.process NodeChildrenChanged fired on path /overseer/queue state SyncConnected
[junit4] 2> 966500 T6255 n:127.0.0.1:50216_y_ow oasc.Overseer$ClusterStateUpdater.run processMessage: queueSize: 1, message = {
[junit4] 2> "operation":"leader",
[junit4] 2> "shard":"shard1",
[junit4] 2> "collection":"collection1",
[junit4] 2> "base_url":"http://127.0.0.1:36422/y_ow",
[junit4] 2> "core":"collection1",
[junit4] 2> "state":"active"} current state version: 7
[junit4] 2> 966556 T6226 oas.SolrTestCaseJ4.writeCoreProperties Writing core.properties file to /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/cores/collection1
[junit4] 2> 966557 T6226 oasc.AbstractFullDistribZkTestBase.createJettys create jetty 3 in directory /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001
[junit4] 2> 966558 T6226 oejs.Server.doStart jetty-9.2.10.v20150310
[junit4] 2> 966559 T6226 oejsh.ContextHandler.doStart Started o.e.j.s.ServletContextHandler@78346c{/y_ow,null,AVAILABLE}
[junit4] 2> 966560 T6226 oejs.AbstractConnector.doStart Started ServerConnector@771834{HTTP/1.1}{127.0.0.1:37028}
[junit4] 2> 966560 T6226 oejs.Server.doStart Started @967496ms
[junit4] 2> 966560 T6226 oascse.JettySolrRunner$1.lifeCycleStarted Jetty properties: {solr.data.dir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/tempDir-001/jetty3, solrconfig=solrconfig.xml, hostContext=/y_ow, hostPort=35144, coreRootDirectory=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/cores}
[junit4] 2> 966560 T6226 oass.SolrDispatchFilter.init SolrDispatchFilter.init()sun.misc.Launcher$AppClassLoader@e2f2a
[junit4] 2> 966561 T6226 oasc.SolrResourceLoader.<init> new SolrResourceLoader for directory: '/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/'
[junit4] 2> 966570 T6226 oasc.SolrXmlConfig.fromFile Loading container configuration from /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/solr.xml
[junit4] 2> 966573 T6226 oasc.CorePropertiesLocator.<init> Config-defined core root directory: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/cores
[junit4] 2> 966573 T6226 oasc.CoreContainer.<init> New CoreContainer 31207983
[junit4] 2> 966574 T6226 oasc.CoreContainer.load Loading cores into CoreContainer [instanceDir=/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/]
[junit4] 2> 966574 T6226 oasc.CoreContainer.load loading shared library: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/lib
[junit4] 2> 966574 T6226 oasc.SolrResourceLoader.addToClassLoader WARN Can't find (or read) directory to add to classloader: lib (resolved as: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001/shard-3-001/lib).
[junit4] 2> 966579 T6226 oashc.HttpShardHandlerFactory.init created with socketTimeout : 90000,urlScheme : ,connTimeout : 15000,maxConnectionsPerHost : 20,maxConnections : 10000,corePoolSize : 0,maximumPoolSize : 2147483647,maxThreadIdleTime : 5,sizeOfQueue : -1,fairnessPolicy : false,useRetries : false,
[junit4] 2> 966580 T6226 oasu.UpdateShardHandler.<init> Creating UpdateShardHandler HTTP client with params: socketTimeout=340000&connTimeout=45000&retry=true
[junit4] 2> 966580 T6226 oasl.LogWatcher.createWatcher SLF4J impl is org.slf4j.impl.Log4jLoggerFactory
[junit4] 2> 966580 T6226 oasl.LogWatcher.newRegisteredLogWatcher Registering Log L
[...truncated too long message...]
[junit4] 2> 5 /solr/aliases.json
[junit4] 2> 5 /solr/clusterstate.json
[junit4] 2> 4 /solr/configs/conf1
[junit4] 2>
[junit4] 2> Maximum concurrent children watches above limit:
[junit4] 2>
[junit4] 2> 5 /solr/live_nodes
[junit4] 2> 4 /solr/overseer/queue
[junit4] 2> 4 /solr/overseer/collection-queue-work
[junit4] 2>
[junit4] 2> 1034796 T6226 c:control_collection s:shard1 x:collection1 oasc.SocketProxy.close WARN Closing 10 connections to: http://127.0.0.1:41627/y_ow, target: http://127.0.0.1:35724/y_ow
[junit4] 2> 1034797 T6226 c:control_collection s:shard1 x:collection1 oasc.SocketProxy.close WARN Closing 4 connections to: http://127.0.0.1:50216/y_ow, target: http://127.0.0.1:40508/y_ow
[junit4] 2> 1034797 T6226 c:control_collection s:shard1 x:collection1 oasc.SocketProxy.close WARN Closing 6 connections to: http://127.0.0.1:35144/y_ow, target: http://127.0.0.1:37028/y_ow
[junit4] 2> 1034797 T6226 c:control_collection s:shard1 x:collection1 oasc.SocketProxy.close WARN Closing 1 connections to: http://127.0.0.1:36422/y_ow, target: http://127.0.0.1:40120/y_ow
[junit4] 2> NOTE: reproduce with: ant test -Dtestcase=HttpPartitionTest -Dtests.method=test -Dtests.seed=375FC4C9173DBE9E -Dtests.multiplier=3 -Dtests.slow=true -Dtests.locale=ar_SD -Dtests.timezone=Atlantic/Jan_Mayen -Dtests.asserts=true -Dtests.file.encoding=UTF-8
[junit4] FAILURE 74.3s J0 | HttpPartitionTest.test <<<
[junit4] > Throwable #1: java.lang.AssertionError: Didn't see all replicas for shard shard1 in c8n_1x2 come up within 30000 ms! ClusterState: {
[junit4] > "collection1":{
[junit4] > "replicationFactor":"1",
[junit4] > "shards":{
[junit4] > "shard1":{
[junit4] > "range":"80000000-ffffffff",
[junit4] > "state":"active",
[junit4] > "replicas":{"core_node2":{
[junit4] > "core":"collection1",
[junit4] > "base_url":"http://127.0.0.1:36422/y_ow",
[junit4] > "node_name":"127.0.0.1:36422_y_ow",
[junit4] > "state":"active",
[junit4] > "leader":"true"}}},
[junit4] > "shard2":{
[junit4] > "range":"0-7fffffff",
[junit4] > "state":"active",
[junit4] > "replicas":{
[junit4] > "core_node1":{
[junit4] > "core":"collection1",
[junit4] > "base_url":"http://127.0.0.1:41627/y_ow",
[junit4] > "node_name":"127.0.0.1:41627_y_ow",
[junit4] > "state":"active",
[junit4] > "leader":"true"},
[junit4] > "core_node3":{
[junit4] > "core":"collection1",
[junit4] > "base_url":"http://127.0.0.1:35144/y_ow",
[junit4] > "node_name":"127.0.0.1:35144_y_ow",
[junit4] > "state":"active"}}}},
[junit4] > "router":{"name":"compositeId"},
[junit4] > "maxShardsPerNode":"1",
[junit4] > "autoAddReplicas":"false",
[junit4] > "autoCreated":"true"},
[junit4] > "control_collection":{
[junit4] > "replicationFactor":"1",
[junit4] > "shards":{"shard1":{
[junit4] > "range":"80000000-7fffffff",
[junit4] > "state":"active",
[junit4] > "replicas":{"core_node1":{
[junit4] > "core":"collection1",
[junit4] > "base_url":"http://127.0.0.1:50216/y_ow",
[junit4] > "node_name":"127.0.0.1:50216_y_ow",
[junit4] > "state":"active",
[junit4] > "leader":"true"}}}},
[junit4] > "router":{"name":"compositeId"},
[junit4] > "maxShardsPerNode":"1",
[junit4] > "autoAddReplicas":"false",
[junit4] > "autoCreated":"true"},
[junit4] > "c8n_1x2":{
[junit4] > "replicationFactor":"2",
[junit4] > "shards":{"shard1":{
[junit4] > "range":"80000000-7fffffff",
[junit4] > "state":"active",
[junit4] > "replicas":{
[junit4] > "core_node1":{
[junit4] > "core":"c8n_1x2_shard1_replica1",
[junit4] > "base_url":"http://127.0.0.1:36422/y_ow",
[junit4] > "node_name":"127.0.0.1:36422_y_ow",
[junit4] > "state":"recovering"},
[junit4] > "core_node2":{
[junit4] > "core":"c8n_1x2_shard1_replica2",
[junit4] > "base_url":"http://127.0.0.1:35144/y_ow",
[junit4] > "node_name":"127.0.0.1:35144_y_ow",
[junit4] > "state":"active",
[junit4] > "leader":"true"}}}},
[junit4] > "router":{"name":"compositeId"},
[junit4] > "maxShardsPerNode":"1",
[junit4] > "autoAddReplicas":"false"}}
[junit4] > at __randomizedtesting.SeedInfo.seed([375FC4C9173DBE9E:BF0BFB13B9C1D366]:0)
[junit4] > at org.apache.solr.cloud.AbstractFullDistribZkTestBase.ensureAllReplicasAreActive(AbstractFullDistribZkTestBase.java:1920)
[junit4] > at org.apache.solr.cloud.HttpPartitionTest.testRf2(HttpPartitionTest.java:205)
[junit4] > at org.apache.solr.cloud.HttpPartitionTest.test(HttpPartitionTest.java:105)
[junit4] > at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsFixedStatement.callStatement(BaseDistributedSearchTestCase.java:960)
[junit4] > at org.apache.solr.BaseDistributedSearchTestCase$ShardsRepeatRule$ShardsStatement.evaluate(BaseDistributedSearchTestCase.java:935)
[junit4] > at java.lang.Thread.run(Thread.java:745)
[junit4] 2> 1034804 T6226 c:control_collection s:shard1 x:collection1 oas.SolrTestCaseJ4.deleteCore ###deleteCore
[junit4] 2> NOTE: leaving temporary files on disk at: /home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build/solr-core/test/J0/temp/solr.cloud.HttpPartitionTest 375FC4C9173DBE9E-001
[junit4] 2> 74259 T6225 ccr.ThreadLeakControl.checkThreadLeaks WARNING Will linger awaiting termination of 1 leaked thread(s).
[junit4] 2> NOTE: test params are: codec=Asserting(Lucene50): {range_facet_l_dv=Lucene50(blocksize=128), _version_=PostingsFormat(name=Memory doPackFST= true), multiDefault=PostingsFormat(name=MockRandom), a_t=PostingsFormat(name=MockRandom), intDefault=PostingsFormat(name=Memory doPackFST= true), id=Lucene50(blocksize=128), range_facet_i_dv=PostingsFormat(name=MockRandom), text=PostingsFormat(name=Asserting), range_facet_l=PostingsFormat(name=MockRandom), timestamp=PostingsFormat(name=MockRandom)}, docValues:{range_facet_l_dv=DocValuesFormat(name=Asserting), range_facet_i_dv=DocValuesFormat(name=Lucene50), timestamp=DocValuesFormat(name=Lucene50)}, sim=RandomSimilarityProvider(queryNorm=false,coord=yes): {}, locale=ar_SD, timezone=Atlantic/Jan_Mayen
[junit4] 2> NOTE: Linux 3.13.0-49-generic i386/Oracle Corporation 1.8.0_60-ea (32-bit)/cpus=12,threads=1,free=83234120,total=205910016
[junit4] 2> NOTE: All tests run in this JVM: [ZkNodePropsTest, SuggesterFSTTest, HighlighterConfigTest, SimpleCollectionCreateDeleteTest, SchemaVersionSpecificBehaviorTest, PeerSyncTest, ExternalFileFieldSortTest, TestStressLucene, TestJmxIntegration, MBeansHandlerTest, DebugComponentTest, TestJoin, TestSolrConfigHandler, QueryElevationComponentTest, TestReversedWildcardFilterFactory, TestConfigOverlay, TestRandomDVFaceting, TestJettySolrRunner, TestCodecSupport, SyncSliceTest, TestDynamicFieldCollectionResource, TestQuerySenderListener, TestRealTimeGet, ChaosMonkeyNothingIsSafeTest, DefaultValueUpdateProcessorTest, TestRangeQuery, DateFieldTest, DistributedFacetPivotSmallTest, SolrCloudExampleTest, TimeZoneUtilsTest, SpatialRPTFieldTypeTest, TestIBSimilarityFactory, TestMiniSolrCloudCluster, TestInfoStreamLogging, TestStandardQParsers, SharedFSAutoReplicaFailoverTest, TestSolr4Spatial, TermVectorComponentTest, LeaderElectionIntegrationTest, HdfsSyncSliceTest, TestDistributedMissingSort, TestRemoteStreaming, ActionThrottleTest, TestConfig, TestMiniSolrCloudClusterSSL, BadCopyFieldTest, TestBlobHandler, BasicDistributedZkTest, TestBadConfig, TestDynamicLoading, TestTrackingShardHandlerFactory, MultiTermTest, TestRandomFaceting, TestSerializedLuceneMatchVersion, TestMergePolicyConfig, TestBulkSchemaConcurrent, ShardSplitTest, TestRecoveryHdfs, TestSolrConfigHandlerCloud, ClusterStateTest, SimpleMLTQParserTest, QueryResultKeyTest, TestReplicationHandlerBackup, TestLRUStatsCache, HttpPartitionTest]
[junit4] Completed [364/485] on J0 in 75.02s, 1 test, 1 failure <<< FAILURES!
[...truncated 370 lines...]
BUILD FAILED
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:526: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:474: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/build.xml:61: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/extra-targets.xml:39: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/build.xml:229: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/solr/common-build.xml:512: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/common-build.xml:1415: The following error occurred while executing this line:
/home/jenkins/workspace/Lucene-Solr-trunk-Linux/lucene/common-build.xml:973: There were test failures: 485 suites, 1947 tests, 1 failure, 57 ignored (25 assumptions)
Total time: 41 minutes 13 seconds
Build step 'Invoke Ant' marked build as failure
Archiving artifacts
Recording test results
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any